var/home/core/zuul-output/0000755000175000017500000000000015127216431014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015127234530015473 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003671703515127234520017712 0ustar rootrootJan 06 14:36:42 crc systemd[1]: Starting Kubernetes Kubelet... Jan 06 14:36:42 crc restorecon[4706]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:42 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Jan 06 14:36:43 crc restorecon[4706]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Jan 06 14:36:43 crc kubenswrapper[4744]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 06 14:36:43 crc kubenswrapper[4744]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 06 14:36:43 crc kubenswrapper[4744]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 06 14:36:43 crc kubenswrapper[4744]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 06 14:36:43 crc kubenswrapper[4744]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 06 14:36:43 crc kubenswrapper[4744]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.513541 4744 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.516966 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.516993 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.516999 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517005 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517009 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517014 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517017 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517023 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517028 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517033 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517038 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517043 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517048 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517053 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517058 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517063 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517068 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517071 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517077 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517081 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517086 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517090 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517094 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517098 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517102 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517105 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517108 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517112 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517115 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517119 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517123 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517126 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517129 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517133 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517136 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517140 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517143 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517147 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517150 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517154 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517172 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517176 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517180 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517184 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517189 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517193 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517197 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517201 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517205 4744 feature_gate.go:330] unrecognized feature gate: Example Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517209 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517213 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517216 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517219 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517223 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517226 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517230 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517234 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517237 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517241 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517244 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517248 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517251 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517255 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517258 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517264 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517268 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517272 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517275 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517279 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517282 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.517286 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517374 4744 flags.go:64] FLAG: --address="0.0.0.0" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517383 4744 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517393 4744 flags.go:64] FLAG: --anonymous-auth="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517401 4744 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517407 4744 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517412 4744 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517418 4744 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517423 4744 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517429 4744 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517434 4744 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517439 4744 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517444 4744 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517448 4744 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517453 4744 flags.go:64] FLAG: --cgroup-root="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517457 4744 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517461 4744 flags.go:64] FLAG: --client-ca-file="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517465 4744 flags.go:64] FLAG: --cloud-config="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517470 4744 flags.go:64] FLAG: --cloud-provider="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517473 4744 flags.go:64] FLAG: --cluster-dns="[]" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517479 4744 flags.go:64] FLAG: --cluster-domain="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517484 4744 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517488 4744 flags.go:64] FLAG: --config-dir="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517492 4744 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517497 4744 flags.go:64] FLAG: --container-log-max-files="5" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517502 4744 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517506 4744 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517511 4744 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517516 4744 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517520 4744 flags.go:64] FLAG: --contention-profiling="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517524 4744 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517528 4744 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517532 4744 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517537 4744 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517542 4744 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517547 4744 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517551 4744 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517556 4744 flags.go:64] FLAG: --enable-load-reader="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517560 4744 flags.go:64] FLAG: --enable-server="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517564 4744 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517570 4744 flags.go:64] FLAG: --event-burst="100" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517575 4744 flags.go:64] FLAG: --event-qps="50" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517580 4744 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517584 4744 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517589 4744 flags.go:64] FLAG: --eviction-hard="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517594 4744 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517598 4744 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517602 4744 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517606 4744 flags.go:64] FLAG: --eviction-soft="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517610 4744 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517615 4744 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517619 4744 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517623 4744 flags.go:64] FLAG: --experimental-mounter-path="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517628 4744 flags.go:64] FLAG: --fail-cgroupv1="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517632 4744 flags.go:64] FLAG: --fail-swap-on="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517636 4744 flags.go:64] FLAG: --feature-gates="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517641 4744 flags.go:64] FLAG: --file-check-frequency="20s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517646 4744 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517651 4744 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517657 4744 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517662 4744 flags.go:64] FLAG: --healthz-port="10248" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517666 4744 flags.go:64] FLAG: --help="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517671 4744 flags.go:64] FLAG: --hostname-override="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517675 4744 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517680 4744 flags.go:64] FLAG: --http-check-frequency="20s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517685 4744 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517689 4744 flags.go:64] FLAG: --image-credential-provider-config="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517693 4744 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517697 4744 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517701 4744 flags.go:64] FLAG: --image-service-endpoint="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517705 4744 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517710 4744 flags.go:64] FLAG: --kube-api-burst="100" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517714 4744 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517718 4744 flags.go:64] FLAG: --kube-api-qps="50" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517724 4744 flags.go:64] FLAG: --kube-reserved="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517728 4744 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517732 4744 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517737 4744 flags.go:64] FLAG: --kubelet-cgroups="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517740 4744 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517745 4744 flags.go:64] FLAG: --lock-file="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517749 4744 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517753 4744 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517757 4744 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517763 4744 flags.go:64] FLAG: --log-json-split-stream="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517767 4744 flags.go:64] FLAG: --log-text-info-buffer-size="0" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517771 4744 flags.go:64] FLAG: --log-text-split-stream="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517775 4744 flags.go:64] FLAG: --logging-format="text" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517779 4744 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517783 4744 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517787 4744 flags.go:64] FLAG: --manifest-url="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517791 4744 flags.go:64] FLAG: --manifest-url-header="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517797 4744 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517801 4744 flags.go:64] FLAG: --max-open-files="1000000" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517806 4744 flags.go:64] FLAG: --max-pods="110" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517811 4744 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517815 4744 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517820 4744 flags.go:64] FLAG: --memory-manager-policy="None" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517824 4744 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517829 4744 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517835 4744 flags.go:64] FLAG: --node-ip="192.168.126.11" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517840 4744 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517850 4744 flags.go:64] FLAG: --node-status-max-images="50" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517855 4744 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517860 4744 flags.go:64] FLAG: --oom-score-adj="-999" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517864 4744 flags.go:64] FLAG: --pod-cidr="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517868 4744 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517875 4744 flags.go:64] FLAG: --pod-manifest-path="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517879 4744 flags.go:64] FLAG: --pod-max-pids="-1" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517883 4744 flags.go:64] FLAG: --pods-per-core="0" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517887 4744 flags.go:64] FLAG: --port="10250" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517891 4744 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517896 4744 flags.go:64] FLAG: --provider-id="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517899 4744 flags.go:64] FLAG: --qos-reserved="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517904 4744 flags.go:64] FLAG: --read-only-port="10255" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517908 4744 flags.go:64] FLAG: --register-node="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517912 4744 flags.go:64] FLAG: --register-schedulable="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517916 4744 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517923 4744 flags.go:64] FLAG: --registry-burst="10" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517927 4744 flags.go:64] FLAG: --registry-qps="5" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517931 4744 flags.go:64] FLAG: --reserved-cpus="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517935 4744 flags.go:64] FLAG: --reserved-memory="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517940 4744 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517944 4744 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517948 4744 flags.go:64] FLAG: --rotate-certificates="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517952 4744 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517956 4744 flags.go:64] FLAG: --runonce="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517960 4744 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517964 4744 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517968 4744 flags.go:64] FLAG: --seccomp-default="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517973 4744 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517977 4744 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517982 4744 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517986 4744 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517990 4744 flags.go:64] FLAG: --storage-driver-password="root" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517994 4744 flags.go:64] FLAG: --storage-driver-secure="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.517999 4744 flags.go:64] FLAG: --storage-driver-table="stats" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518004 4744 flags.go:64] FLAG: --storage-driver-user="root" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518008 4744 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518012 4744 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518017 4744 flags.go:64] FLAG: --system-cgroups="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518020 4744 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518026 4744 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518031 4744 flags.go:64] FLAG: --tls-cert-file="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518035 4744 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518040 4744 flags.go:64] FLAG: --tls-min-version="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518044 4744 flags.go:64] FLAG: --tls-private-key-file="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518048 4744 flags.go:64] FLAG: --topology-manager-policy="none" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518052 4744 flags.go:64] FLAG: --topology-manager-policy-options="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518057 4744 flags.go:64] FLAG: --topology-manager-scope="container" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518061 4744 flags.go:64] FLAG: --v="2" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518066 4744 flags.go:64] FLAG: --version="false" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518072 4744 flags.go:64] FLAG: --vmodule="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518077 4744 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518081 4744 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518186 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518192 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518196 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518201 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518205 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518209 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518213 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518217 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518226 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518230 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518234 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518242 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518245 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518249 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518252 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518256 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518259 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518263 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518267 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518272 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518276 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518281 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518285 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518289 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518293 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518297 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518300 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518304 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518308 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518312 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518316 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518320 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518323 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518327 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518330 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518333 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518337 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518340 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518344 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518347 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518354 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518358 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518362 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518367 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518371 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518377 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518381 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518386 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518389 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518394 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518398 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518402 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518406 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518409 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518413 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518417 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518421 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518425 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518428 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518432 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518436 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518440 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518443 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518447 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518450 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518454 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518458 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518461 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518465 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518468 4744 feature_gate.go:330] unrecognized feature gate: Example Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.518472 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.518483 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.530612 4744 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.530674 4744 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530861 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530874 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530883 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530892 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530900 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530907 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530915 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530923 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530931 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530942 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530955 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530964 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530974 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530985 4744 feature_gate.go:330] unrecognized feature gate: Example Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.530995 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531004 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531012 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531020 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531028 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531037 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531044 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531052 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531060 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531068 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531076 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531087 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531097 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531105 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531114 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531124 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531133 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531142 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531150 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531181 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531189 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531198 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531206 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531214 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531222 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531230 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531238 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531246 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531254 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531261 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531269 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531280 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531290 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531298 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531306 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531314 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531322 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531330 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531337 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531345 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531353 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531362 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531370 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531379 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531387 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531395 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531403 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531411 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531419 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531427 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531435 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531443 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531451 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531459 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531467 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531474 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531482 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.531496 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531747 4744 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531762 4744 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531770 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531779 4744 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531788 4744 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531795 4744 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531803 4744 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531811 4744 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531820 4744 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531828 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531836 4744 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531844 4744 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531852 4744 feature_gate.go:330] unrecognized feature gate: PlatformOperators Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531860 4744 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531868 4744 feature_gate.go:330] unrecognized feature gate: PinnedImages Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531876 4744 feature_gate.go:330] unrecognized feature gate: OVNObservability Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531884 4744 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531892 4744 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531900 4744 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531909 4744 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531916 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531925 4744 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531932 4744 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531943 4744 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531953 4744 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531963 4744 feature_gate.go:330] unrecognized feature gate: SignatureStores Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531972 4744 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531981 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.531989 4744 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532022 4744 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532032 4744 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532040 4744 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532048 4744 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532058 4744 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532068 4744 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532077 4744 feature_gate.go:330] unrecognized feature gate: Example Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532085 4744 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532093 4744 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532102 4744 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532110 4744 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532118 4744 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532126 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532134 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532142 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532150 4744 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532189 4744 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532200 4744 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532210 4744 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532218 4744 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532226 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532233 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532243 4744 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532251 4744 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532259 4744 feature_gate.go:330] unrecognized feature gate: GatewayAPI Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532266 4744 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532274 4744 feature_gate.go:330] unrecognized feature gate: InsightsConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532282 4744 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532291 4744 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532299 4744 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532309 4744 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532318 4744 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532327 4744 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532335 4744 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532343 4744 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532351 4744 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532358 4744 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532366 4744 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532374 4744 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532382 4744 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532389 4744 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.532397 4744 feature_gate.go:330] unrecognized feature gate: NewOLM Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.532408 4744 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.532681 4744 server.go:940] "Client rotation is on, will bootstrap in background" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.537007 4744 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.537133 4744 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.537993 4744 server.go:997] "Starting client certificate rotation" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.538029 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.538387 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-28 16:26:44.810881719 +0000 UTC Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.538585 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.545149 4744 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.547101 4744 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.22:6443: connect: connection refused" logger="UnhandledError" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.547706 4744 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.561599 4744 log.go:25] "Validated CRI v1 runtime API" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.575964 4744 log.go:25] "Validated CRI v1 image API" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.577668 4744 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.580428 4744 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-01-06-14-31-50-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.580472 4744 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.609589 4744 manager.go:217] Machine: {Timestamp:2026-01-06 14:36:43.607266228 +0000 UTC m=+0.234732596 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:87748cd0-7b55-4658-b5b9-16b00729a3a5 BootID:ba2b6707-2ee9-4315-912c-dcfa1a01ee80 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:4f:bc:96 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:4f:bc:96 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:2f:0b:3d Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:aa:13:f4 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:21:85:c7 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:fc:46:e2 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:9a:ca:45:8e:91:d3 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:0e:27:5a:ff:77:2c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.610389 4744 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.610721 4744 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.611377 4744 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.611853 4744 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.612111 4744 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.612776 4744 topology_manager.go:138] "Creating topology manager with none policy" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.612931 4744 container_manager_linux.go:303] "Creating device plugin manager" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.613442 4744 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.613698 4744 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.614245 4744 state_mem.go:36] "Initialized new in-memory state store" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.614590 4744 server.go:1245] "Using root directory" path="/var/lib/kubelet" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.615835 4744 kubelet.go:418] "Attempting to sync node with API server" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.615997 4744 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.616248 4744 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.616458 4744 kubelet.go:324] "Adding apiserver pod source" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.616602 4744 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.618507 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.618624 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.22:6443: connect: connection refused" logger="UnhandledError" Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.618803 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.618959 4744 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.619048 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.22:6443: connect: connection refused" logger="UnhandledError" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.619548 4744 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.620697 4744 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621565 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621622 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621642 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621662 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621691 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621708 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621727 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621748 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621765 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621779 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621799 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.621813 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.622424 4744 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.623256 4744 server.go:1280] "Started kubelet" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.623264 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.623644 4744 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.623666 4744 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.624503 4744 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 06 14:36:43 crc systemd[1]: Started Kubernetes Kubelet. Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.626972 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.22:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18882b1080ca9174 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-06 14:36:43.62314994 +0000 UTC m=+0.250616298,LastTimestamp:2026-01-06 14:36:43.62314994 +0000 UTC m=+0.250616298,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.628814 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.628874 4744 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.628999 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 04:32:37.216509732 +0000 UTC Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.629096 4744 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 277h55m53.587418727s for next certificate rotation Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.629240 4744 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.630397 4744 server.go:460] "Adding debug handlers to kubelet server" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.630434 4744 volume_manager.go:287] "The desired_state_of_world populator starts" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.630466 4744 volume_manager.go:289] "Starting Kubelet Volume Manager" Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.630512 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" interval="200ms" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.630631 4744 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.638450 4744 factory.go:55] Registering systemd factory Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.638759 4744 factory.go:221] Registration of the systemd container factory successfully Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.638849 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.638994 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.22:6443: connect: connection refused" logger="UnhandledError" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.639836 4744 factory.go:153] Registering CRI-O factory Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.639883 4744 factory.go:221] Registration of the crio container factory successfully Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.639980 4744 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.640014 4744 factory.go:103] Registering Raw factory Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.640035 4744 manager.go:1196] Started watching for new ooms in manager Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.640953 4744 manager.go:319] Starting recovery of all containers Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.653869 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654017 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654436 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654466 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654487 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654507 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654526 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654550 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654574 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654674 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654693 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654714 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654733 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654758 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654778 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654797 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654893 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654912 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654930 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654947 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654967 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.654989 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655009 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655030 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655093 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655114 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655144 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655191 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655255 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655318 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655420 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655445 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655556 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655577 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655603 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655628 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655656 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655684 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655706 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655733 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655758 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655783 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655809 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655850 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655878 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655960 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.655988 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656013 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656199 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656233 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656260 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656284 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656368 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656400 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656429 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656458 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656515 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656536 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656555 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656573 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656592 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656610 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656628 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656647 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656666 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656685 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656704 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656721 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656742 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656775 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656792 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656815 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656846 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656865 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656882 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656901 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656918 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656938 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656955 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656973 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.656994 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657011 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657032 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657050 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657068 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657085 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657103 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657123 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657141 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657186 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657208 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657389 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657412 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657432 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657450 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657468 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657553 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657571 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657590 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657608 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657629 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657648 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657666 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657683 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657757 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657779 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657798 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657819 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657838 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657860 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657880 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657899 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657921 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657940 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657958 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657977 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.657994 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658025 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658043 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658060 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658080 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658100 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658117 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658136 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658153 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658196 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658216 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658235 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658255 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658273 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658290 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658310 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658335 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658353 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658380 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658399 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658418 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658436 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658455 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658473 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658492 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658510 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658526 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658544 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658563 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658581 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658600 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658630 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658648 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658668 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658686 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658749 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658766 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658784 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658804 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658824 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658850 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658868 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658886 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658904 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658924 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658942 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658960 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658978 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.658999 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659016 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659033 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659051 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659067 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659085 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659105 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659122 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659144 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659185 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.659606 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.661830 4744 manager.go:324] Recovery completed Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.662070 4744 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.662112 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.662212 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.662232 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.662251 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.662395 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.662414 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.662432 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.662619 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665426 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665558 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665620 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665691 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665739 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665782 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665817 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665848 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665890 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665929 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.665977 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666012 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666047 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666090 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666126 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666194 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666243 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666280 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666328 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666361 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666393 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666432 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666463 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666503 4744 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666527 4744 reconstruct.go:97] "Volume reconstruction finished" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.666547 4744 reconciler.go:26] "Reconciler: start to sync state" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.675030 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.676749 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.676789 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.676798 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.678477 4744 cpu_manager.go:225] "Starting CPU manager" policy="none" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.678540 4744 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.678565 4744 state_mem.go:36] "Initialized new in-memory state store" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.690353 4744 policy_none.go:49] "None policy: Start" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.691575 4744 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.691605 4744 state_mem.go:35] "Initializing new in-memory state store" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.707832 4744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.709763 4744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.709810 4744 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.709832 4744 kubelet.go:2335] "Starting kubelet main sync loop" Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.709883 4744 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 06 14:36:43 crc kubenswrapper[4744]: W0106 14:36:43.711121 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.712283 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.22:6443: connect: connection refused" logger="UnhandledError" Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.729798 4744 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.752337 4744 manager.go:334] "Starting Device Plugin manager" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.752389 4744 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.752405 4744 server.go:79] "Starting device plugin registration server" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.752809 4744 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.752829 4744 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.752990 4744 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.753077 4744 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.753087 4744 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.765134 4744 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.810990 4744 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc"] Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.811201 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.812807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.812878 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.812898 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.813144 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.813456 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.813539 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.814465 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.814501 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.814513 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.814540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.814558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.814567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.814667 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.814699 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.814843 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.816061 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.816130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.816151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.816782 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.816836 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.816855 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.817054 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.817180 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.817215 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818147 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818192 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818315 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818454 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818497 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818563 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818630 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.818967 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.819015 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.819024 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.819210 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.819252 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.819688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.819730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.819748 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.820707 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.820754 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.820772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.831471 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" interval="400ms" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.853125 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.854944 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.854988 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.855022 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.855052 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 06 14:36:43 crc kubenswrapper[4744]: E0106 14:36:43.855634 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.22:6443: connect: connection refused" node="crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869326 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869401 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869596 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869652 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869733 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869789 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869889 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869927 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.869985 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.870032 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.870077 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.870124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.870202 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971392 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971462 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971492 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971523 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971554 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971582 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971610 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971611 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971643 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971645 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971727 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971637 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971698 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971823 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971825 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971733 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971926 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971976 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.971993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.972071 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.972114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.972153 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.972202 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.972265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.972281 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:36:43 crc kubenswrapper[4744]: I0106 14:36:43.972338 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.055968 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.059030 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.059107 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.059136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.059193 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 06 14:36:44 crc kubenswrapper[4744]: E0106 14:36:44.060133 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.22:6443: connect: connection refused" node="crc" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.139482 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.148749 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.170212 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:36:44 crc kubenswrapper[4744]: W0106 14:36:44.178639 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-9b0baa822aea793caae7723a2ac244847013274d67c000633347cbc8e7b8cbf4 WatchSource:0}: Error finding container 9b0baa822aea793caae7723a2ac244847013274d67c000633347cbc8e7b8cbf4: Status 404 returned error can't find the container with id 9b0baa822aea793caae7723a2ac244847013274d67c000633347cbc8e7b8cbf4 Jan 06 14:36:44 crc kubenswrapper[4744]: W0106 14:36:44.183365 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-c4a8ce47bc9f8cc5f93e1dbaa0759c4c443318293fac095ea0a04169ed51aec4 WatchSource:0}: Error finding container c4a8ce47bc9f8cc5f93e1dbaa0759c4c443318293fac095ea0a04169ed51aec4: Status 404 returned error can't find the container with id c4a8ce47bc9f8cc5f93e1dbaa0759c4c443318293fac095ea0a04169ed51aec4 Jan 06 14:36:44 crc kubenswrapper[4744]: W0106 14:36:44.201232 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-7ca6f6954c05e2dda064358e2d944b7ebedd06df8d57ca5e0faad796c05d503a WatchSource:0}: Error finding container 7ca6f6954c05e2dda064358e2d944b7ebedd06df8d57ca5e0faad796c05d503a: Status 404 returned error can't find the container with id 7ca6f6954c05e2dda064358e2d944b7ebedd06df8d57ca5e0faad796c05d503a Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.201302 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.209480 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Jan 06 14:36:44 crc kubenswrapper[4744]: W0106 14:36:44.224142 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-1a133038905d4759b6368c495f8a56be80d8b3b9f40f4604bdac37dd8715ee4e WatchSource:0}: Error finding container 1a133038905d4759b6368c495f8a56be80d8b3b9f40f4604bdac37dd8715ee4e: Status 404 returned error can't find the container with id 1a133038905d4759b6368c495f8a56be80d8b3b9f40f4604bdac37dd8715ee4e Jan 06 14:36:44 crc kubenswrapper[4744]: E0106 14:36:44.232829 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" interval="800ms" Jan 06 14:36:44 crc kubenswrapper[4744]: W0106 14:36:44.236012 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-61689c331bd7ad1271a2c7bd41c1965f4e371bd777524299cbf901f75de8a581 WatchSource:0}: Error finding container 61689c331bd7ad1271a2c7bd41c1965f4e371bd777524299cbf901f75de8a581: Status 404 returned error can't find the container with id 61689c331bd7ad1271a2c7bd41c1965f4e371bd777524299cbf901f75de8a581 Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.461222 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.462453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.462493 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.462504 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.462531 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 06 14:36:44 crc kubenswrapper[4744]: E0106 14:36:44.463027 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.22:6443: connect: connection refused" node="crc" Jan 06 14:36:44 crc kubenswrapper[4744]: W0106 14:36:44.523866 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:44 crc kubenswrapper[4744]: E0106 14:36:44.523951 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.22:6443: connect: connection refused" logger="UnhandledError" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.623932 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.716195 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a50ee7ce34ab8458b2d63585d172bfa39cc97f974fd6144e403ca0b36fc1be53"} Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.716369 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c4a8ce47bc9f8cc5f93e1dbaa0759c4c443318293fac095ea0a04169ed51aec4"} Jan 06 14:36:44 crc kubenswrapper[4744]: W0106 14:36:44.716548 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:44 crc kubenswrapper[4744]: E0106 14:36:44.716682 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.22:6443: connect: connection refused" logger="UnhandledError" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.718233 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371" exitCode=0 Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.718384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371"} Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.718462 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9b0baa822aea793caae7723a2ac244847013274d67c000633347cbc8e7b8cbf4"} Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.718661 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.720065 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0" exitCode=0 Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.720215 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0"} Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.720290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"61689c331bd7ad1271a2c7bd41c1965f4e371bd777524299cbf901f75de8a581"} Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.720483 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.720925 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.720968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.720983 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.721859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.721892 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.721900 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.721999 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="c2eac205efc889c0f6727a673653c906f0b47a206b69a140dde0f9e89e8dd8bd" exitCode=0 Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.722090 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"c2eac205efc889c0f6727a673653c906f0b47a206b69a140dde0f9e89e8dd8bd"} Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.722134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1a133038905d4759b6368c495f8a56be80d8b3b9f40f4604bdac37dd8715ee4e"} Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.722258 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.723151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.723204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.723223 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.723343 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.726322 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.726404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.726462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.729043 4744 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="8d0ef1b2410b948046e0c8249ead80d57bec6757bf3c2356f2ec23898aefda2c" exitCode=0 Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.729103 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"8d0ef1b2410b948046e0c8249ead80d57bec6757bf3c2356f2ec23898aefda2c"} Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.729149 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7ca6f6954c05e2dda064358e2d944b7ebedd06df8d57ca5e0faad796c05d503a"} Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.729281 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.730084 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.730136 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:44 crc kubenswrapper[4744]: I0106 14:36:44.730184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:44 crc kubenswrapper[4744]: W0106 14:36:44.759261 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:44 crc kubenswrapper[4744]: E0106 14:36:44.759356 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.22:6443: connect: connection refused" logger="UnhandledError" Jan 06 14:36:44 crc kubenswrapper[4744]: W0106 14:36:44.905935 4744 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.22:6443: connect: connection refused Jan 06 14:36:44 crc kubenswrapper[4744]: E0106 14:36:44.906017 4744 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.22:6443: connect: connection refused" logger="UnhandledError" Jan 06 14:36:45 crc kubenswrapper[4744]: E0106 14:36:45.033619 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" interval="1.6s" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.264112 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.265231 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.265263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.265274 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.265312 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 06 14:36:45 crc kubenswrapper[4744]: E0106 14:36:45.265652 4744 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.22:6443: connect: connection refused" node="crc" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.717024 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.734222 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"114f039bc7fab51c84600395897226678214a9205c40dac0837805b1352eb6b1"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.734434 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.735762 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.735817 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.735835 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.736798 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c9f139be18e48af594c1934b4f29ea33f926e126001d3a484289278cc5e28c2d"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.736878 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f2c20e9df0711bb17d52ccd57d1c0181826328ca9c9a93b26b580cc2e00efa26"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.736909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"58850b54a6e6ab043a7fd522771e32faf410b4a759b30a39c383ebdea8a7a57f"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.736827 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.737986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.738049 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.738068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.739596 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3deaa537d947fe5bd69ecf6272557e287b1b35c269888facd7bc32c978c22f72"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.739644 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.739654 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1d8a731da82dba89ac5fba517b79a25b133393d406da4de29b95dc96eeb03570"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.739803 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"04e29aabdb20bb91ef985171b456e40f5750a06c32c64c612471bbf3842a1b1a"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.740862 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.740928 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.740952 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.743269 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.743320 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.743342 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.743359 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.746147 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0" exitCode=0 Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.746221 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0"} Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.746460 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.747645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.747704 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:45 crc kubenswrapper[4744]: I0106 14:36:45.747723 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.051313 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.060337 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.752702 4744 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8" exitCode=0 Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.752782 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8"} Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.752994 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.754408 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.754471 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.754490 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.758143 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b"} Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.758260 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.758348 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.758387 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.758358 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.760093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.760109 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.760237 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.760256 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.760109 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.760340 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.760359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.760204 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.760468 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.866502 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.868132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.868216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.868235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:46 crc kubenswrapper[4744]: I0106 14:36:46.868267 4744 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.455944 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.767324 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.767755 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"53c03ad22b668987eee3d7c892a52e7514380b25c4d50b00d5132dc2f50f5b05"} Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.767801 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"352eaf5271bf661543010442e975a35b3283d695d34f0a84694c948bccf19686"} Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.767824 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d5d7358920ba690c8b8a58dd0432dc4641c5817b6e3287a5f4d73528e59529b7"} Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.768010 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.768058 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.768202 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.768776 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.768822 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.768840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.769826 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.769880 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.769897 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.794637 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.795006 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.796456 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.796511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:47 crc kubenswrapper[4744]: I0106 14:36:47.796526 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.598546 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.776812 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"da48209b52d3f93a9a1d5013595a1fa7a0bbfe01c37e58b0636b24538b45f1a2"} Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.776909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3315d32d031bd5b7ec0eb48702e428d06ccb64c5613c1281d80a7d7349824fce"} Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.776967 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.776996 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.776988 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.779265 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.779324 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.779351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.779866 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.779911 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.779921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.779975 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.779930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:48 crc kubenswrapper[4744]: I0106 14:36:48.780005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:49 crc kubenswrapper[4744]: I0106 14:36:49.541386 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:49 crc kubenswrapper[4744]: I0106 14:36:49.779812 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:49 crc kubenswrapper[4744]: I0106 14:36:49.779825 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:49 crc kubenswrapper[4744]: I0106 14:36:49.781625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:49 crc kubenswrapper[4744]: I0106 14:36:49.781677 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:49 crc kubenswrapper[4744]: I0106 14:36:49.781695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:49 crc kubenswrapper[4744]: I0106 14:36:49.782231 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:49 crc kubenswrapper[4744]: I0106 14:36:49.782411 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:49 crc kubenswrapper[4744]: I0106 14:36:49.782549 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.051831 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.052717 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.054536 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.054609 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.054633 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.178427 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.178650 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.180409 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.180495 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.180599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.512134 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.785410 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.787055 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.787123 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:51 crc kubenswrapper[4744]: I0106 14:36:51.787142 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:53 crc kubenswrapper[4744]: E0106 14:36:53.765278 4744 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.208047 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.208114 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.588621 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.588696 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.626620 4744 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.651349 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.651558 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.652914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.652966 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.652985 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.658383 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:36:55 crc kubenswrapper[4744]: E0106 14:36:55.721909 4744 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.795485 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.796643 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.796679 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:36:55 crc kubenswrapper[4744]: I0106 14:36:55.796693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:36:56 crc kubenswrapper[4744]: I0106 14:36:56.456393 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 06 14:36:56 crc kubenswrapper[4744]: I0106 14:36:56.456482 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 06 14:36:56 crc kubenswrapper[4744]: I0106 14:36:56.464532 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Jan 06 14:36:56 crc kubenswrapper[4744]: I0106 14:36:56.464615 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Jan 06 14:36:57 crc kubenswrapper[4744]: I0106 14:36:57.468848 4744 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]log ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]etcd ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-api-request-count-filter ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-startkubeinformers ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/generic-apiserver-start-informers ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/priority-and-fairness-config-consumer ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/priority-and-fairness-filter ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-apiextensions-informers ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-apiextensions-controllers ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/crd-informer-synced ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-system-namespaces-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-cluster-authentication-info-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-legacy-token-tracking-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-service-ip-repair-controllers ok Jan 06 14:36:57 crc kubenswrapper[4744]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Jan 06 14:36:57 crc kubenswrapper[4744]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/priority-and-fairness-config-producer ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/bootstrap-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/start-kube-aggregator-informers ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/apiservice-status-local-available-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/apiservice-status-remote-available-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/apiservice-registration-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/apiservice-wait-for-first-sync ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/apiservice-discovery-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/kube-apiserver-autoregistration ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]autoregister-completion ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/apiservice-openapi-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: [+]poststarthook/apiservice-openapiv3-controller ok Jan 06 14:36:57 crc kubenswrapper[4744]: livez check failed Jan 06 14:36:57 crc kubenswrapper[4744]: I0106 14:36:57.468931 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:36:58 crc kubenswrapper[4744]: I0106 14:36:58.651587 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 14:36:58 crc kubenswrapper[4744]: I0106 14:36:58.651671 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 06 14:36:59 crc kubenswrapper[4744]: I0106 14:36:59.929580 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 06 14:36:59 crc kubenswrapper[4744]: I0106 14:36:59.951468 4744 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 06 14:36:59 crc kubenswrapper[4744]: I0106 14:36:59.981026 4744 csr.go:261] certificate signing request csr-vxhd9 is approved, waiting to be issued Jan 06 14:36:59 crc kubenswrapper[4744]: I0106 14:36:59.991695 4744 csr.go:257] certificate signing request csr-vxhd9 is issued Jan 06 14:37:00 crc kubenswrapper[4744]: I0106 14:37:00.993701 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-06 14:31:59 +0000 UTC, rotation deadline is 2026-10-19 05:05:04.241022999 +0000 UTC Jan 06 14:37:00 crc kubenswrapper[4744]: I0106 14:37:00.995079 4744 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6854h28m3.245971369s for next certificate rotation Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.231559 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.231763 4744 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.232908 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.232936 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.232949 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.244347 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.476377 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.480389 4744 trace.go:236] Trace[961511583]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Jan-2026 14:36:47.373) (total time: 14106ms): Jan 06 14:37:01 crc kubenswrapper[4744]: Trace[961511583]: ---"Objects listed" error: 14106ms (14:37:01.480) Jan 06 14:37:01 crc kubenswrapper[4744]: Trace[961511583]: [14.106289059s] [14.106289059s] END Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.480439 4744 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.482741 4744 trace.go:236] Trace[1748548849]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Jan-2026 14:36:46.483) (total time: 14998ms): Jan 06 14:37:01 crc kubenswrapper[4744]: Trace[1748548849]: ---"Objects listed" error: 14998ms (14:37:01.482) Jan 06 14:37:01 crc kubenswrapper[4744]: Trace[1748548849]: [14.998879483s] [14.998879483s] END Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.482788 4744 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.485990 4744 trace.go:236] Trace[226811278]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Jan-2026 14:36:47.874) (total time: 13611ms): Jan 06 14:37:01 crc kubenswrapper[4744]: Trace[226811278]: ---"Objects listed" error: 13611ms (14:37:01.485) Jan 06 14:37:01 crc kubenswrapper[4744]: Trace[226811278]: [13.611521219s] [13.611521219s] END Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.486356 4744 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.486425 4744 trace.go:236] Trace[1681867580]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Jan-2026 14:36:47.361) (total time: 14125ms): Jan 06 14:37:01 crc kubenswrapper[4744]: Trace[1681867580]: ---"Objects listed" error: 14125ms (14:37:01.486) Jan 06 14:37:01 crc kubenswrapper[4744]: Trace[1681867580]: [14.125330304s] [14.125330304s] END Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.486885 4744 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.487226 4744 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.501043 4744 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.501402 4744 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.503257 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.503306 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.503323 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.503345 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.503356 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:01Z","lastTransitionTime":"2026-01-06T14:37:01Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.529475 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ba2b6707-2ee9-4315-912c-dcfa1a01ee80\\\",\\\"systemUUID\\\":\\\"87748cd0-7b55-4658-b5b9-16b00729a3a5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.534737 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.534792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.534807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.534831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.534854 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:01Z","lastTransitionTime":"2026-01-06T14:37:01Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.547736 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ba2b6707-2ee9-4315-912c-dcfa1a01ee80\\\",\\\"systemUUID\\\":\\\"87748cd0-7b55-4658-b5b9-16b00729a3a5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.554884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.554931 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.554945 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.554971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.554987 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:01Z","lastTransitionTime":"2026-01-06T14:37:01Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.577488 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ba2b6707-2ee9-4315-912c-dcfa1a01ee80\\\",\\\"systemUUID\\\":\\\"87748cd0-7b55-4658-b5b9-16b00729a3a5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.583634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.583688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.583696 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.583715 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.583727 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:01Z","lastTransitionTime":"2026-01-06T14:37:01Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.596499 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ba2b6707-2ee9-4315-912c-dcfa1a01ee80\\\",\\\"systemUUID\\\":\\\"87748cd0-7b55-4658-b5b9-16b00729a3a5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.601807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.601840 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.601847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.601865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.601876 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:01Z","lastTransitionTime":"2026-01-06T14:37:01Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.613848 4744 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ba2b6707-2ee9-4315-912c-dcfa1a01ee80\\\",\\\"systemUUID\\\":\\\"87748cd0-7b55-4658-b5b9-16b00729a3a5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.613993 4744 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.615434 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.615464 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.615473 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.615491 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.615501 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:01Z","lastTransitionTime":"2026-01-06T14:37:01Z","reason":"KubeletNotReady","message":"[container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?, CSINode is not yet initialized]"} Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.629973 4744 apiserver.go:52] "Watching apiserver" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.634932 4744 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.635292 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.635696 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.635810 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.635822 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.636267 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.636267 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.636337 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.636413 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.636468 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.636514 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.638088 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.638124 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.639899 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.641198 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.641228 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.641791 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.642060 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.642547 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.642778 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.690788 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.717886 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.717927 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.717937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.717952 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.717960 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:01Z","lastTransitionTime":"2026-01-06T14:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.731330 4744 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.734319 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.764075 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-8mz44"] Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.764389 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-8mz44" Jan 06 14:37:01 crc kubenswrapper[4744]: W0106 14:37:01.770206 4744 reflector.go:561] object-"openshift-dns"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.770252 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:01 crc kubenswrapper[4744]: W0106 14:37:01.770205 4744 reflector.go:561] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": failed to list *v1.Secret: secrets "node-resolver-dockercfg-kz9s7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.770289 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"node-resolver-dockercfg-kz9s7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"node-resolver-dockercfg-kz9s7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.770271 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: W0106 14:37:01.770337 4744 reflector.go:561] object-"openshift-dns"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.770388 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.785848 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.788931 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.788987 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789007 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789025 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789040 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789178 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789204 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789222 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789241 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789257 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789272 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789287 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789303 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789319 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789334 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789349 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789368 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789387 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789405 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789841 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789485 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789876 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789490 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789898 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789917 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789934 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789588 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789603 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789627 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789639 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789827 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.789858 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.789957 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:02.289935562 +0000 UTC m=+18.917401880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790303 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790344 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790365 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790385 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790404 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790438 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790461 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790481 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790500 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790524 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790541 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790560 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790585 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790606 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790628 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790653 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790672 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790696 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790715 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790737 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790754 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790774 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790790 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790807 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790825 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790840 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790859 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790880 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790898 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790916 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790937 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790957 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790976 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.790995 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791014 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791031 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791050 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791068 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791085 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791103 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791129 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791147 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791197 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791217 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791235 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791257 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791280 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791299 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791315 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791333 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791352 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791370 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791388 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791405 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791425 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791446 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791471 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791490 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791510 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791528 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791582 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791600 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791619 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791638 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791656 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791675 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791692 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791709 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791727 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791743 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791761 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791779 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791798 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791837 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791856 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791877 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791913 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791930 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791946 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791963 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.791985 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792003 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792021 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792037 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792053 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792069 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792086 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792102 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792120 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792137 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792153 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792186 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792202 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792221 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792239 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792273 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792291 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792308 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792325 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792342 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792359 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792377 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792396 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792413 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792431 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792447 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792468 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792485 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792504 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792521 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792540 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792565 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792584 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792602 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792622 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792639 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792657 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792674 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792691 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792709 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792728 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792748 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792765 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792784 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792806 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792824 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792842 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792858 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792875 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792893 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792910 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792929 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792946 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792963 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.792984 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793003 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793023 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793042 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793063 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793082 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793104 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793123 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793142 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793716 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793741 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793764 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793787 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793808 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793832 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793854 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793877 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793896 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793916 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793934 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793953 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793972 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.793992 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794010 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794030 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794050 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794070 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794089 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794108 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794126 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794145 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794197 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794216 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794234 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794294 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794322 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794374 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794407 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794455 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794481 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794504 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794527 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794575 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794699 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794713 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.810041 4744 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.819868 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.840950 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841016 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841034 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841052 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841080 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:01Z","lastTransitionTime":"2026-01-06T14:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794403 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794560 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.794767 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.853059 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.795148 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.795241 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.795293 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.795464 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.795507 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.795572 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.795656 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.795771 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.795845 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796149 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796223 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796472 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796500 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796623 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796637 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796742 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796770 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796810 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.796979 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.798013 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.799033 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.799981 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.800006 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.800211 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.800248 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.800431 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.800438 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.800709 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.801011 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.801063 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.801249 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.801417 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.801606 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.801758 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.801775 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.802006 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.802060 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.853776 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.802293 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.802680 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.802913 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.802996 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.803193 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.803406 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.803572 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.803649 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.803966 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804078 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804098 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804183 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804236 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804381 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804469 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804599 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804708 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804746 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.804928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.805005 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.810692 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.811044 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.802312 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.814664 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.818060 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.818181 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.818426 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.818585 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.818840 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.802323 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.820208 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.820365 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.820611 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.820683 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.824469 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.824856 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.824861 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.824908 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.825126 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.825375 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.825497 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.825976 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.826577 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.826633 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.827110 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.827148 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.852288 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.827282 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.827362 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.827454 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.835936 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.836122 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.836517 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.836703 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.836848 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.837053 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.837135 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.837195 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.837314 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.837328 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.837478 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.837507 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.837987 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.840408 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.840412 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.840442 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.840741 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.840824 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841028 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841345 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841434 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841449 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.841608 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.842079 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.842304 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.842311 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.842526 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.843322 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.844174 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.845208 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.845629 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.845835 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.846835 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.846972 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.847335 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.847405 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.847509 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.847601 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.847762 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.848186 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.848330 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.848639 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.848684 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.849177 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.850731 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.851052 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.851066 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.851113 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.851130 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.851419 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.852056 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.852305 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.852409 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.852444 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.852680 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.852925 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.854024 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.854684 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.855181 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.855663 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.855866 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.852805 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.855999 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.856201 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.856346 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.857172 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.857402 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.857559 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.857802 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.858340 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.858643 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.858488 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.860073 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.859940 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.860153 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.860368 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.860852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.862377 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.863100 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:02.363078289 +0000 UTC m=+18.990544607 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.863113 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.863284 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:02.363262724 +0000 UTC m=+18.990729042 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.863677 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.863938 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.864597 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.865232 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.865513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.866092 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.866175 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.866555 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.867383 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.867579 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.867644 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.868005 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.868426 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.869324 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.869975 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.869994 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.870007 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.870064 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:02.370047773 +0000 UTC m=+18.997514091 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.872192 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.877105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.877244 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.879455 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.879481 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.879492 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:01 crc kubenswrapper[4744]: E0106 14:37:01.879534 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:02.379519322 +0000 UTC m=+19.006985640 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.884417 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.887602 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.887846 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.897937 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.897993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.897944 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898051 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/38f0ac6a-e5ca-48e4-960f-371ac91370a6-hosts-file\") pod \"node-resolver-8mz44\" (UID: \"38f0ac6a-e5ca-48e4-960f-371ac91370a6\") " pod="openshift-dns/node-resolver-8mz44" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898092 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898110 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q57n4\" (UniqueName: \"kubernetes.io/projected/38f0ac6a-e5ca-48e4-960f-371ac91370a6-kube-api-access-q57n4\") pod \"node-resolver-8mz44\" (UID: \"38f0ac6a-e5ca-48e4-960f-371ac91370a6\") " pod="openshift-dns/node-resolver-8mz44" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898235 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898247 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898257 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898267 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898278 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898289 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898299 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898307 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898317 4744 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898326 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898336 4744 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898346 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898357 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898366 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898377 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898385 4744 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898394 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898403 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898412 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898421 4744 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898430 4744 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898439 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898448 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898451 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898461 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898472 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898483 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898494 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898504 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898514 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898523 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898532 4744 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898542 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898552 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898562 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898570 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898579 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898588 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898597 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898606 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898615 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898624 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898633 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898642 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898654 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898664 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898675 4744 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898684 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898694 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898705 4744 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898714 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898724 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898734 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898743 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898752 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898763 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898771 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898780 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898789 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898798 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898806 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898815 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898823 4744 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898831 4744 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898841 4744 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898849 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898858 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898877 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898885 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898894 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898904 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898912 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898921 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898929 4744 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898937 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898945 4744 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898954 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898962 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898973 4744 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898981 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.898991 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899000 4744 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899008 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899016 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899025 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899034 4744 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899043 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899050 4744 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899059 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899067 4744 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899077 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899066 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899085 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899139 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899152 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899187 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899198 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899208 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899226 4744 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899239 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899250 4744 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899259 4744 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899269 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899279 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899289 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899298 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899307 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899318 4744 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899328 4744 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899338 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899348 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899357 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899367 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899378 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899388 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899399 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899411 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899422 4744 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899432 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899441 4744 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899453 4744 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899462 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899473 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899482 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899495 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899505 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899515 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899524 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899533 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899545 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899513 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899560 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899745 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899766 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899784 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899798 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899810 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899822 4744 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899835 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899850 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899862 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899894 4744 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899907 4744 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899919 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899931 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899943 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899956 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899968 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899979 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.899992 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900004 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900017 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900029 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900042 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900055 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900070 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900082 4744 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900095 4744 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900109 4744 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900122 4744 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900134 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900146 4744 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900177 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900187 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900196 4744 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900205 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900214 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900222 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900232 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900241 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900249 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900258 4744 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900267 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900275 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900285 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900295 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900304 4744 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900313 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900323 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900334 4744 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900344 4744 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900354 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900364 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900373 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900383 4744 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900391 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900400 4744 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900408 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900418 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900426 4744 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.900436 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.902745 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b" exitCode=255 Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.902807 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b"} Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.910653 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.912755 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.915595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.920291 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.922826 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.924191 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.937524 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.943637 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.943699 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.943714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.943730 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.943741 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:01Z","lastTransitionTime":"2026-01-06T14:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.949767 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.950318 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.957521 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.961072 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.965215 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.965758 4744 scope.go:117] "RemoveContainer" containerID="30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.971490 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 06 14:37:01 crc kubenswrapper[4744]: W0106 14:37:01.972197 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-ada56a47f2dd573c393a19cf41c286ca89aed5d484cfc6765555d30f67c0fbca WatchSource:0}: Error finding container ada56a47f2dd573c393a19cf41c286ca89aed5d484cfc6765555d30f67c0fbca: Status 404 returned error can't find the container with id ada56a47f2dd573c393a19cf41c286ca89aed5d484cfc6765555d30f67c0fbca Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.974951 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:01 crc kubenswrapper[4744]: I0106 14:37:01.987211 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8mz44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38f0ac6a-e5ca-48e4-960f-371ac91370a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q57n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8mz44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.001019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/38f0ac6a-e5ca-48e4-960f-371ac91370a6-hosts-file\") pod \"node-resolver-8mz44\" (UID: \"38f0ac6a-e5ca-48e4-960f-371ac91370a6\") " pod="openshift-dns/node-resolver-8mz44" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.001069 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q57n4\" (UniqueName: \"kubernetes.io/projected/38f0ac6a-e5ca-48e4-960f-371ac91370a6-kube-api-access-q57n4\") pod \"node-resolver-8mz44\" (UID: \"38f0ac6a-e5ca-48e4-960f-371ac91370a6\") " pod="openshift-dns/node-resolver-8mz44" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.001123 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.001136 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.001146 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.001172 4744 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.001538 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/38f0ac6a-e5ca-48e4-960f-371ac91370a6-hosts-file\") pod \"node-resolver-8mz44\" (UID: \"38f0ac6a-e5ca-48e4-960f-371ac91370a6\") " pod="openshift-dns/node-resolver-8mz44" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.002054 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.012886 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.022581 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.031772 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.035461 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.047806 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.050560 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.050584 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.050592 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.050610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.050622 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.063013 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85d85475-e337-4034-8a0c-1f6e873d9476\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0106 14:36:56.077814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0106 14:36:56.079825 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2496083149/tls.crt::/tmp/serving-cert-2496083149/tls.key\\\\\\\"\\\\nI0106 14:37:01.581363 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0106 14:37:01.586267 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0106 14:37:01.586293 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0106 14:37:01.586321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0106 14:37:01.586327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0106 14:37:01.602990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0106 14:37:01.603028 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603034 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0106 14:37:01.603043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0106 14:37:01.603047 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0106 14:37:01.603051 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0106 14:37:01.603408 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0106 14:37:01.616083 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.077498 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.089814 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8mz44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38f0ac6a-e5ca-48e4-960f-371ac91370a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q57n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8mz44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.154424 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.154458 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.154495 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.154511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.154521 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.187611 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-rm6lj"] Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.187967 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-jcmms"] Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.188122 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kc5zk"] Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.188378 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.188743 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-jsxwz"] Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.188986 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.189169 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.190002 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.196947 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.197203 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.197635 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.197870 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198077 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198152 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198183 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198205 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198324 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198330 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198415 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198455 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198501 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198506 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198569 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198654 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.198673 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.199801 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.200142 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.203083 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.217246 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.225591 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1639b17c-035d-4955-a3b6-ac33889e1c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rm6lj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.244428 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"827dde89-8f82-4aeb-b89b-a7ae2a2125dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://352eaf5271bf661543010442e975a35b3283d695d34f0a84694c948bccf19686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c03ad22b668987eee3d7c892a52e7514380b25c4d50b00d5132dc2f50f5b05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3315d32d031bd5b7ec0eb48702e428d06ccb64c5613c1281d80a7d7349824fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da48209b52d3f93a9a1d5013595a1fa7a0bbfe01c37e58b0636b24538b45f1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5d7358920ba690c8b8a58dd0432dc4641c5817b6e3287a5f4d73528e59529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.255598 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.256553 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.256585 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.256597 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.256613 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.256623 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.283026 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.295282 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.302750 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.302863 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:03.302837752 +0000 UTC m=+19.930304070 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.302917 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-systemd-units\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.302956 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-ovn\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.302977 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-script-lib\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.302998 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-hostroot\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303017 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snkfs\" (UniqueName: \"kubernetes.io/projected/1639b17c-035d-4955-a3b6-ac33889e1c89-kube-api-access-snkfs\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303057 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-var-lib-kubelet\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1639b17c-035d-4955-a3b6-ac33889e1c89-proxy-tls\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303105 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-etc-openvswitch\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-openvswitch\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-bin\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303188 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1639b17c-035d-4955-a3b6-ac33889e1c89-mcd-auth-proxy-config\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-cni-dir\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303226 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-netns\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303257 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e4e53714-2c82-4edb-aafc-cc9f3890a645-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303277 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-os-release\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303313 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0cbc5afb-14db-45d6-b564-93a47a9e8123-cni-binary-copy\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303353 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-socket-dir-parent\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303378 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-etc-kubernetes\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303430 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-var-lib-cni-bin\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303450 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-var-lib-openvswitch\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303484 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-system-cni-dir\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303509 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-os-release\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njg45\" (UniqueName: \"kubernetes.io/projected/0cbc5afb-14db-45d6-b564-93a47a9e8123-kube-api-access-njg45\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303548 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-systemd\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303568 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-log-socket\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303610 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-system-cni-dir\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303631 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-run-netns\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303647 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-node-log\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303704 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-netd\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303725 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9ec2daf2-1809-448d-9891-795e5dff0696-ovn-node-metrics-cert\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303755 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-conf-dir\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303773 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-run-multus-certs\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303793 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303813 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-var-lib-cni-multus\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303861 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-cnibin\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303940 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-config\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.303986 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e4e53714-2c82-4edb-aafc-cc9f3890a645-cni-binary-copy\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-run-k8s-cni-cncf-io\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-daemon-config\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304067 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnls9\" (UniqueName: \"kubernetes.io/projected/e4e53714-2c82-4edb-aafc-cc9f3890a645-kube-api-access-hnls9\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304101 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-kubelet\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/1639b17c-035d-4955-a3b6-ac33889e1c89-rootfs\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304154 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-slash\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-ovn-kubernetes\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304235 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-env-overrides\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304252 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp2kl\" (UniqueName: \"kubernetes.io/projected/9ec2daf2-1809-448d-9891-795e5dff0696-kube-api-access-gp2kl\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.304274 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-cnibin\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.306323 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85d85475-e337-4034-8a0c-1f6e873d9476\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0106 14:36:56.077814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0106 14:36:56.079825 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2496083149/tls.crt::/tmp/serving-cert-2496083149/tls.key\\\\\\\"\\\\nI0106 14:37:01.581363 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0106 14:37:01.586267 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0106 14:37:01.586293 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0106 14:37:01.586321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0106 14:37:01.586327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0106 14:37:01.602990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0106 14:37:01.603028 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603034 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0106 14:37:01.603043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0106 14:37:01.603047 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0106 14:37:01.603051 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0106 14:37:01.603408 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0106 14:37:01.616083 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.315803 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.326921 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8mz44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38f0ac6a-e5ca-48e4-960f-371ac91370a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q57n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8mz44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.338350 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.352284 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.358848 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.358879 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.358890 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.358905 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.358919 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.373293 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.389830 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.400123 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1639b17c-035d-4955-a3b6-ac33889e1c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rm6lj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404687 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/1639b17c-035d-4955-a3b6-ac33889e1c89-rootfs\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404730 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnls9\" (UniqueName: \"kubernetes.io/projected/e4e53714-2c82-4edb-aafc-cc9f3890a645-kube-api-access-hnls9\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404749 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-kubelet\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404767 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-cnibin\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404781 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-slash\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404796 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-ovn-kubernetes\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404802 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/1639b17c-035d-4955-a3b6-ac33889e1c89-rootfs\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404813 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-env-overrides\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404871 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp2kl\" (UniqueName: \"kubernetes.io/projected/9ec2daf2-1809-448d-9891-795e5dff0696-kube-api-access-gp2kl\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-hostroot\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-ovn-kubernetes\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404870 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-cnibin\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404910 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-systemd-units\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404976 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-slash\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404928 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-systemd-units\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404891 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-kubelet\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.404952 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-hostroot\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405011 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-ovn\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405038 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-script-lib\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405051 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-ovn\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405073 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405098 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snkfs\" (UniqueName: \"kubernetes.io/projected/1639b17c-035d-4955-a3b6-ac33889e1c89-kube-api-access-snkfs\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405124 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-var-lib-kubelet\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405146 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-openvswitch\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-bin\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405214 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1639b17c-035d-4955-a3b6-ac33889e1c89-proxy-tls\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405232 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-etc-openvswitch\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405264 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-bin\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405277 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-var-lib-kubelet\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405290 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1639b17c-035d-4955-a3b6-ac33889e1c89-mcd-auth-proxy-config\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.405361 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405371 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-etc-openvswitch\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-cni-dir\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405344 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-openvswitch\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.405408 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.405476 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:03.405453164 +0000 UTC m=+20.032919662 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405451 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-netns\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405931 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-etc-kubernetes\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405479 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-netns\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405651 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-script-lib\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.405474 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406038 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406083 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e4e53714-2c82-4edb-aafc-cc9f3890a645-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-os-release\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406124 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0cbc5afb-14db-45d6-b564-93a47a9e8123-cni-binary-copy\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406048 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-env-overrides\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.406047 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406184 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1639b17c-035d-4955-a3b6-ac33889e1c89-mcd-auth-proxy-config\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406202 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-socket-dir-parent\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406144 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-socket-dir-parent\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406232 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-system-cni-dir\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406253 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-os-release\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.406257 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:03.406230365 +0000 UTC m=+20.033696873 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.405583 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-cni-dir\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406266 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-system-cni-dir\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.406137 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406302 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-var-lib-cni-bin\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406005 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-etc-kubernetes\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.406347 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:03.406333178 +0000 UTC m=+20.033799696 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406372 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-var-lib-openvswitch\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406398 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-var-lib-cni-bin\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406402 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-log-socket\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406428 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-log-socket\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-os-release\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406492 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njg45\" (UniqueName: \"kubernetes.io/projected/0cbc5afb-14db-45d6-b564-93a47a9e8123-kube-api-access-njg45\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-systemd\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406464 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-var-lib-openvswitch\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-netd\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406579 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9ec2daf2-1809-448d-9891-795e5dff0696-ovn-node-metrics-cert\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-systemd\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-system-cni-dir\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406580 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-os-release\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406641 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-run-netns\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406709 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e4e53714-2c82-4edb-aafc-cc9f3890a645-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406737 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-run-netns\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406751 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-node-log\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406762 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-system-cni-dir\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406765 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-netd\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-node-log\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.406797 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.407150 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-tuning-conf-dir\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.407362 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/0cbc5afb-14db-45d6-b564-93a47a9e8123-cni-binary-copy\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408213 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-var-lib-cni-multus\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408241 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-conf-dir\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-run-multus-certs\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408277 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408293 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-cnibin\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408315 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408330 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-config\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e4e53714-2c82-4edb-aafc-cc9f3890a645-cni-binary-copy\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408364 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-run-k8s-cni-cncf-io\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408378 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-daemon-config\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408841 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-daemon-config\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408882 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-var-lib-cni-multus\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-multus-conf-dir\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.408931 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-run-multus-certs\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.409008 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.409025 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.409036 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.409079 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:03.40906574 +0000 UTC m=+20.036532058 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.409106 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e4e53714-2c82-4edb-aafc-cc9f3890a645-cnibin\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.409129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.409526 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-config\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.409897 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e4e53714-2c82-4edb-aafc-cc9f3890a645-cni-binary-copy\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.409929 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/0cbc5afb-14db-45d6-b564-93a47a9e8123-host-run-k8s-cni-cncf-io\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.410609 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9ec2daf2-1809-448d-9891-795e5dff0696-ovn-node-metrics-cert\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.411633 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1639b17c-035d-4955-a3b6-ac33889e1c89-proxy-tls\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.422143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp2kl\" (UniqueName: \"kubernetes.io/projected/9ec2daf2-1809-448d-9891-795e5dff0696-kube-api-access-gp2kl\") pod \"ovnkube-node-kc5zk\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.422638 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njg45\" (UniqueName: \"kubernetes.io/projected/0cbc5afb-14db-45d6-b564-93a47a9e8123-kube-api-access-njg45\") pod \"multus-jcmms\" (UID: \"0cbc5afb-14db-45d6-b564-93a47a9e8123\") " pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.425245 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85d85475-e337-4034-8a0c-1f6e873d9476\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0106 14:36:56.077814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0106 14:36:56.079825 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2496083149/tls.crt::/tmp/serving-cert-2496083149/tls.key\\\\\\\"\\\\nI0106 14:37:01.581363 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0106 14:37:01.586267 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0106 14:37:01.586293 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0106 14:37:01.586321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0106 14:37:01.586327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0106 14:37:01.602990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0106 14:37:01.603028 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603034 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0106 14:37:01.603043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0106 14:37:01.603047 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0106 14:37:01.603051 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0106 14:37:01.603408 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0106 14:37:01.616083 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.426203 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnls9\" (UniqueName: \"kubernetes.io/projected/e4e53714-2c82-4edb-aafc-cc9f3890a645-kube-api-access-hnls9\") pod \"multus-additional-cni-plugins-jsxwz\" (UID: \"e4e53714-2c82-4edb-aafc-cc9f3890a645\") " pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.427716 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snkfs\" (UniqueName: \"kubernetes.io/projected/1639b17c-035d-4955-a3b6-ac33889e1c89-kube-api-access-snkfs\") pod \"machine-config-daemon-rm6lj\" (UID: \"1639b17c-035d-4955-a3b6-ac33889e1c89\") " pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.443585 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcmms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbc5afb-14db-45d6-b564-93a47a9e8123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcmms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.462379 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.462410 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.462419 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.462432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.462442 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.468271 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.469209 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ec2daf2-1809-448d-9891-795e5dff0696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kc5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.499083 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"827dde89-8f82-4aeb-b89b-a7ae2a2125dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://352eaf5271bf661543010442e975a35b3283d695d34f0a84694c948bccf19686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c03ad22b668987eee3d7c892a52e7514380b25c4d50b00d5132dc2f50f5b05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3315d32d031bd5b7ec0eb48702e428d06ccb64c5613c1281d80a7d7349824fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da48209b52d3f93a9a1d5013595a1fa7a0bbfe01c37e58b0636b24538b45f1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5d7358920ba690c8b8a58dd0432dc4641c5817b6e3287a5f4d73528e59529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.504990 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.513246 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jcmms" Jan 06 14:37:02 crc kubenswrapper[4744]: W0106 14:37:02.522438 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cbc5afb_14db_45d6_b564_93a47a9e8123.slice/crio-6b3403d46585b24e89438855193edad23c01964f0e425551c867d50a87afede1 WatchSource:0}: Error finding container 6b3403d46585b24e89438855193edad23c01964f0e425551c867d50a87afede1: Status 404 returned error can't find the container with id 6b3403d46585b24e89438855193edad23c01964f0e425551c867d50a87afede1 Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.524663 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.538492 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.543153 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.545292 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.550503 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8mz44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38f0ac6a-e5ca-48e4-960f-371ac91370a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q57n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8mz44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.569720 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.570036 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.570096 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.570176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.570255 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.589694 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e53714-2c82-4edb-aafc-cc9f3890a645\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jsxwz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.619501 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.637766 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8mz44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38f0ac6a-e5ca-48e4-960f-371ac91370a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q57n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8mz44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.667366 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e53714-2c82-4edb-aafc-cc9f3890a645\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jsxwz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.675054 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.675181 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.675246 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.675337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.675407 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.685987 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.697633 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.710583 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.710714 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.710595 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:02 crc kubenswrapper[4744]: E0106 14:37:02.710794 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.712454 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.726707 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.739662 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1639b17c-035d-4955-a3b6-ac33889e1c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rm6lj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.751558 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85d85475-e337-4034-8a0c-1f6e873d9476\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0106 14:36:56.077814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0106 14:36:56.079825 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2496083149/tls.crt::/tmp/serving-cert-2496083149/tls.key\\\\\\\"\\\\nI0106 14:37:01.581363 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0106 14:37:01.586267 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0106 14:37:01.586293 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0106 14:37:01.586321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0106 14:37:01.586327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0106 14:37:01.602990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0106 14:37:01.603028 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603034 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0106 14:37:01.603043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0106 14:37:01.603047 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0106 14:37:01.603051 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0106 14:37:01.603408 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0106 14:37:01.616083 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.763513 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcmms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbc5afb-14db-45d6-b564-93a47a9e8123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcmms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.778239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.778285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.778296 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.778313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.778324 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.801503 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ec2daf2-1809-448d-9891-795e5dff0696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kc5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.842733 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"827dde89-8f82-4aeb-b89b-a7ae2a2125dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://352eaf5271bf661543010442e975a35b3283d695d34f0a84694c948bccf19686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c03ad22b668987eee3d7c892a52e7514380b25c4d50b00d5132dc2f50f5b05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3315d32d031bd5b7ec0eb48702e428d06ccb64c5613c1281d80a7d7349824fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da48209b52d3f93a9a1d5013595a1fa7a0bbfe01c37e58b0636b24538b45f1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5d7358920ba690c8b8a58dd0432dc4641c5817b6e3287a5f4d73528e59529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.890742 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.893623 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.893674 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.893686 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.893706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.893716 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.906580 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ada56a47f2dd573c393a19cf41c286ca89aed5d484cfc6765555d30f67c0fbca"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.908141 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"74d004b017a439658b55e271008782d9546f9354325db0bfc9e13d50687d3dff"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.908197 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9139268eceb6ad608f755b46b344215aa2b19fc7070269c82782807834b9d6e7"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.911740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"a05f2bceb66e3248444e47316944a9a8778ecd9d1da91e8710cbc6487a76ada9"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.911827 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"f0db91145c265a1792dad8ccd75d8c9a1942a698f2bf63e4a4bdd7ae0b74f422"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.911849 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"fea117dd3f50771be49cb44868f1ac88b663a47cacd1ce897bd5531e83549ae0"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.913623 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ab4c4ec93e1910e491b227ac42dd9ec2ad693d8209d6f77e5a2490f4b2a4194d"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.913739 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"99d6f4b7437fa48f5e9e9f9624c65ac716464efbd807e9f3bd9a5ddeb4a598e3"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.914348 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e07470e09f3ccd42eb78c61e069f00718ed2fa755d76a2430914d717e3aa1149"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.916030 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.918109 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.918560 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.919522 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ec2daf2-1809-448d-9891-795e5dff0696" containerID="b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789" exitCode=0 Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.919590 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.919632 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerStarted","Data":"945ebaac79098f8b6e0344435149a9517e1c4714ebaa58c4f7d970c741f422a8"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.920901 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" event={"ID":"e4e53714-2c82-4edb-aafc-cc9f3890a645","Type":"ContainerStarted","Data":"f82c892dfe397df490f78d9320524fc0a4108283ffa2640ce82d88aea2a4f2d4"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.920957 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" event={"ID":"e4e53714-2c82-4edb-aafc-cc9f3890a645","Type":"ContainerStarted","Data":"7615a819b274d0a7467e8158b743f1df8c741736494ed357afd03a7d63953bb4"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.923001 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcmms" event={"ID":"0cbc5afb-14db-45d6-b564-93a47a9e8123","Type":"ContainerStarted","Data":"d54efa8efe73152bbe8099225f978ca4b236f6799956aa12b40a9e99d5394847"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.923086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcmms" event={"ID":"0cbc5afb-14db-45d6-b564-93a47a9e8123","Type":"ContainerStarted","Data":"6b3403d46585b24e89438855193edad23c01964f0e425551c867d50a87afede1"} Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.926338 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.930897 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.956866 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.999467 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.999520 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.999535 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.999556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:02 crc kubenswrapper[4744]: I0106 14:37:02.999569 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:02Z","lastTransitionTime":"2026-01-06T14:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.004215 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:02Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.006772 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.014134 4744 projected.go:288] Couldn't get configMap openshift-dns/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.014239 4744 projected.go:194] Error preparing data for projected volume kube-api-access-q57n4 for pod openshift-dns/node-resolver-8mz44: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.014350 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/38f0ac6a-e5ca-48e4-960f-371ac91370a6-kube-api-access-q57n4 podName:38f0ac6a-e5ca-48e4-960f-371ac91370a6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:03.514331961 +0000 UTC m=+20.141798279 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-q57n4" (UniqueName: "kubernetes.io/projected/38f0ac6a-e5ca-48e4-960f-371ac91370a6-kube-api-access-q57n4") pod "node-resolver-8mz44" (UID: "38f0ac6a-e5ca-48e4-960f-371ac91370a6") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.059204 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1639b17c-035d-4955-a3b6-ac33889e1c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rm6lj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.096778 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d004b017a439658b55e271008782d9546f9354325db0bfc9e13d50687d3dff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.104683 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.104755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.104769 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.104793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.104804 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:03Z","lastTransitionTime":"2026-01-06T14:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.140412 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcmms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbc5afb-14db-45d6-b564-93a47a9e8123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcmms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.147204 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.202697 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ec2daf2-1809-448d-9891-795e5dff0696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kc5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.207044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.207094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.207111 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.207148 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.207190 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:03Z","lastTransitionTime":"2026-01-06T14:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.248117 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85d85475-e337-4034-8a0c-1f6e873d9476\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0106 14:36:56.077814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0106 14:36:56.079825 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2496083149/tls.crt::/tmp/serving-cert-2496083149/tls.key\\\\\\\"\\\\nI0106 14:37:01.581363 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0106 14:37:01.586267 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0106 14:37:01.586293 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0106 14:37:01.586321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0106 14:37:01.586327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0106 14:37:01.602990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0106 14:37:01.603028 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603034 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0106 14:37:01.603043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0106 14:37:01.603047 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0106 14:37:01.603051 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0106 14:37:01.603408 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0106 14:37:01.616083 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.282049 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"827dde89-8f82-4aeb-b89b-a7ae2a2125dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://352eaf5271bf661543010442e975a35b3283d695d34f0a84694c948bccf19686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c03ad22b668987eee3d7c892a52e7514380b25c4d50b00d5132dc2f50f5b05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3315d32d031bd5b7ec0eb48702e428d06ccb64c5613c1281d80a7d7349824fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da48209b52d3f93a9a1d5013595a1fa7a0bbfe01c37e58b0636b24538b45f1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5d7358920ba690c8b8a58dd0432dc4641c5817b6e3287a5f4d73528e59529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.291788 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.309667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.309695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.309705 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.309721 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.309731 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:03Z","lastTransitionTime":"2026-01-06T14:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.319267 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.319526 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:05.319509559 +0000 UTC m=+21.946975867 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.335059 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.378720 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.412523 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.412558 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.412567 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.412581 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.412592 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:03Z","lastTransitionTime":"2026-01-06T14:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.419961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.420014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.420045 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.420065 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420185 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420231 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:05.420218131 +0000 UTC m=+22.047684449 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420539 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420563 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420577 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420603 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:05.420596351 +0000 UTC m=+22.048062669 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420646 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420654 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420660 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420679 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:05.420672963 +0000 UTC m=+22.048139281 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420707 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.420728 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:05.420723404 +0000 UTC m=+22.048189722 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.431348 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8mz44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38f0ac6a-e5ca-48e4-960f-371ac91370a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q57n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8mz44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.472599 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e53714-2c82-4edb-aafc-cc9f3890a645\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jsxwz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.496660 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://74d004b017a439658b55e271008782d9546f9354325db0bfc9e13d50687d3dff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.516110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.516143 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.516151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.516203 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.516213 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:03Z","lastTransitionTime":"2026-01-06T14:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.521567 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q57n4\" (UniqueName: \"kubernetes.io/projected/38f0ac6a-e5ca-48e4-960f-371ac91370a6-kube-api-access-q57n4\") pod \"node-resolver-8mz44\" (UID: \"38f0ac6a-e5ca-48e4-960f-371ac91370a6\") " pod="openshift-dns/node-resolver-8mz44" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.535189 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q57n4\" (UniqueName: \"kubernetes.io/projected/38f0ac6a-e5ca-48e4-960f-371ac91370a6-kube-api-access-q57n4\") pod \"node-resolver-8mz44\" (UID: \"38f0ac6a-e5ca-48e4-960f-371ac91370a6\") " pod="openshift-dns/node-resolver-8mz44" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.538505 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.539521 4744 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 06 14:37:03 crc kubenswrapper[4744]: W0106 14:37:03.539850 4744 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 06 14:37:03 crc kubenswrapper[4744]: W0106 14:37:03.540121 4744 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 06 14:37:03 crc kubenswrapper[4744]: W0106 14:37:03.540527 4744 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.574173 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-8mz44" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.599370 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.619452 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.619481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.619489 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.619502 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.619512 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:03Z","lastTransitionTime":"2026-01-06T14:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.624791 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ab4c4ec93e1910e491b227ac42dd9ec2ad693d8209d6f77e5a2490f4b2a4194d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99d6f4b7437fa48f5e9e9f9624c65ac716464efbd807e9f3bd9a5ddeb4a598e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.661408 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1639b17c-035d-4955-a3b6-ac33889e1c89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a05f2bceb66e3248444e47316944a9a8778ecd9d1da91e8710cbc6487a76ada9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0db91145c265a1792dad8ccd75d8c9a1942a698f2bf63e4a4bdd7ae0b74f422\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snkfs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rm6lj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.710237 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.710261 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"85d85475-e337-4034-8a0c-1f6e873d9476\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0106 14:36:56.077814 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0106 14:36:56.079825 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2496083149/tls.crt::/tmp/serving-cert-2496083149/tls.key\\\\\\\"\\\\nI0106 14:37:01.581363 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0106 14:37:01.586267 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0106 14:37:01.586293 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0106 14:37:01.586321 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0106 14:37:01.586327 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0106 14:37:01.602990 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0106 14:37:01.603028 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603034 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0106 14:37:01.603039 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0106 14:37:01.603043 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0106 14:37:01.603047 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0106 14:37:01.603051 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0106 14:37:01.603408 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0106 14:37:01.616083 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: E0106 14:37:03.710384 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.716715 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.717355 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.718866 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.719643 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.720923 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.723636 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.724520 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.725948 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.726908 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.728462 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.728511 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.728529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.728554 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.728581 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:03Z","lastTransitionTime":"2026-01-06T14:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.733833 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.734900 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.737145 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.737985 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.738664 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.739784 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.741839 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jcmms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0cbc5afb-14db-45d6-b564-93a47a9e8123\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d54efa8efe73152bbe8099225f978ca4b236f6799956aa12b40a9e99d5394847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-njg45\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jcmms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.741961 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.742633 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.743577 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.744397 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.745062 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.746695 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.747524 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.748726 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.750723 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.751860 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.753284 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.756336 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.756888 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.758719 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.759832 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.760808 4744 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.760930 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.763347 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.763869 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.764739 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.766447 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.767534 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.768785 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.769964 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.770743 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.772691 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.774100 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.777277 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.779149 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.782390 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.783121 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.783008 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ec2daf2-1809-448d-9891-795e5dff0696\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gp2kl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-kc5zk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.784344 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.785279 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.788115 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.788821 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.790713 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.791479 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.792128 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.794374 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.821148 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"827dde89-8f82-4aeb-b89b-a7ae2a2125dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://352eaf5271bf661543010442e975a35b3283d695d34f0a84694c948bccf19686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c03ad22b668987eee3d7c892a52e7514380b25c4d50b00d5132dc2f50f5b05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3315d32d031bd5b7ec0eb48702e428d06ccb64c5613c1281d80a7d7349824fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da48209b52d3f93a9a1d5013595a1fa7a0bbfe01c37e58b0636b24538b45f1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5d7358920ba690c8b8a58dd0432dc4641c5817b6e3287a5f4d73528e59529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.831102 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.831132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.831140 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.831176 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.831190 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:03Z","lastTransitionTime":"2026-01-06T14:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.854921 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.896222 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.929387 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerStarted","Data":"e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.929450 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerStarted","Data":"60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.929466 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerStarted","Data":"feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.931751 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4e53714-2c82-4edb-aafc-cc9f3890a645" containerID="f82c892dfe397df490f78d9320524fc0a4108283ffa2640ce82d88aea2a4f2d4" exitCode=0 Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.931820 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" event={"ID":"e4e53714-2c82-4edb-aafc-cc9f3890a645","Type":"ContainerDied","Data":"f82c892dfe397df490f78d9320524fc0a4108283ffa2640ce82d88aea2a4f2d4"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.932905 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.932942 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.932956 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.932976 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.933004 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:03Z","lastTransitionTime":"2026-01-06T14:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.933485 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-8mz44" event={"ID":"38f0ac6a-e5ca-48e4-960f-371ac91370a6","Type":"ContainerStarted","Data":"dfa44164b4413881d1aea93937730b23a4d540acb24f409952c5de407ce82676"} Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.937891 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-8mz44" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"38f0ac6a-e5ca-48e4-960f-371ac91370a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q57n4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-8mz44\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:03 crc kubenswrapper[4744]: I0106 14:37:03.985644 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4e53714-2c82-4edb-aafc-cc9f3890a645\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f82c892dfe397df490f78d9320524fc0a4108283ffa2640ce82d88aea2a4f2d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:37:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hnls9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:37:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-jsxwz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:03Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.022878 4744 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"827dde89-8f82-4aeb-b89b-a7ae2a2125dd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-06T14:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://352eaf5271bf661543010442e975a35b3283d695d34f0a84694c948bccf19686\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53c03ad22b668987eee3d7c892a52e7514380b25c4d50b00d5132dc2f50f5b05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3315d32d031bd5b7ec0eb48702e428d06ccb64c5613c1281d80a7d7349824fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da48209b52d3f93a9a1d5013595a1fa7a0bbfe01c37e58b0636b24538b45f1a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5d7358920ba690c8b8a58dd0432dc4641c5817b6e3287a5f4d73528e59529b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-06T14:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4149d911a6635072740e285287d44b447e084ea8d86fbdda982940ba7c50d0f0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9154530dfd91c52cb5fea8bf13b5a86854132104d4acb34cdb896f314dbe50d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd0d92428593de501b818a00dacfdb69d587f8c15d44b63285f2d39a68267de8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-06T14:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-06T14:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-06T14:36:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-06T14:37:04Z is after 2025-08-24T17:21:41Z" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.036598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.036634 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.036645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.036662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.036675 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.120319 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-98zj4"] Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.120705 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.139312 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.139346 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.139359 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.139375 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.139386 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.147277 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.167278 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.187209 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.211091 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.227183 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zjqt\" (UniqueName: \"kubernetes.io/projected/9fc78f49-6e10-4a01-a37b-8de049da84ae-kube-api-access-6zjqt\") pod \"node-ca-98zj4\" (UID: \"9fc78f49-6e10-4a01-a37b-8de049da84ae\") " pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.227278 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9fc78f49-6e10-4a01-a37b-8de049da84ae-serviceca\") pod \"node-ca-98zj4\" (UID: \"9fc78f49-6e10-4a01-a37b-8de049da84ae\") " pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.227307 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fc78f49-6e10-4a01-a37b-8de049da84ae-host\") pod \"node-ca-98zj4\" (UID: \"9fc78f49-6e10-4a01-a37b-8de049da84ae\") " pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.243693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.243731 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.243744 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.243758 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.243768 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.328691 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9fc78f49-6e10-4a01-a37b-8de049da84ae-serviceca\") pod \"node-ca-98zj4\" (UID: \"9fc78f49-6e10-4a01-a37b-8de049da84ae\") " pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.329110 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fc78f49-6e10-4a01-a37b-8de049da84ae-host\") pod \"node-ca-98zj4\" (UID: \"9fc78f49-6e10-4a01-a37b-8de049da84ae\") " pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.329200 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zjqt\" (UniqueName: \"kubernetes.io/projected/9fc78f49-6e10-4a01-a37b-8de049da84ae-kube-api-access-6zjqt\") pod \"node-ca-98zj4\" (UID: \"9fc78f49-6e10-4a01-a37b-8de049da84ae\") " pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.329281 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fc78f49-6e10-4a01-a37b-8de049da84ae-host\") pod \"node-ca-98zj4\" (UID: \"9fc78f49-6e10-4a01-a37b-8de049da84ae\") " pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.330137 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9fc78f49-6e10-4a01-a37b-8de049da84ae-serviceca\") pod \"node-ca-98zj4\" (UID: \"9fc78f49-6e10-4a01-a37b-8de049da84ae\") " pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.346569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.346612 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.346625 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.346645 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.346658 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.370496 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zjqt\" (UniqueName: \"kubernetes.io/projected/9fc78f49-6e10-4a01-a37b-8de049da84ae-kube-api-access-6zjqt\") pod \"node-ca-98zj4\" (UID: \"9fc78f49-6e10-4a01-a37b-8de049da84ae\") " pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.434612 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podStartSLOduration=3.434587758 podStartE2EDuration="3.434587758s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:04.43427176 +0000 UTC m=+21.061738088" watchObservedRunningTime="2026-01-06 14:37:04.434587758 +0000 UTC m=+21.062054076" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.449703 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.449754 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.449766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.449785 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.449798 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.451423 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js"] Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.451951 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.488713 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.501594 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-98zj4" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.502421 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-f9m2v"] Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.502835 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:04 crc kubenswrapper[4744]: E0106 14:37:04.502893 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f9m2v" podUID="4e45835e-b6fd-4d25-ad1a-c5d9b9221cde" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.507399 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 06 14:37:04 crc kubenswrapper[4744]: W0106 14:37:04.519562 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fc78f49_6e10_4a01_a37b_8de049da84ae.slice/crio-611f75dedf2a5fdc1b13a5ff2c4c2bc5eb90fd2ede8fba702935d11bd703a307 WatchSource:0}: Error finding container 611f75dedf2a5fdc1b13a5ff2c4c2bc5eb90fd2ede8fba702935d11bd703a307: Status 404 returned error can't find the container with id 611f75dedf2a5fdc1b13a5ff2c4c2bc5eb90fd2ede8fba702935d11bd703a307 Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.530531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfgfr\" (UniqueName: \"kubernetes.io/projected/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-kube-api-access-nfgfr\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.530581 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.530607 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.530668 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.552227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.552443 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.552455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.552476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.552487 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.607877 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-jcmms" podStartSLOduration=3.607847722 podStartE2EDuration="3.607847722s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:04.559363125 +0000 UTC m=+21.186829433" watchObservedRunningTime="2026-01-06 14:37:04.607847722 +0000 UTC m=+21.235314030" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.631761 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.631812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm8fs\" (UniqueName: \"kubernetes.io/projected/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-kube-api-access-gm8fs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.631864 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.631958 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfgfr\" (UniqueName: \"kubernetes.io/projected/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-kube-api-access-nfgfr\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.631981 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.632002 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.633116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-env-overrides\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.633645 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.644116 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.656777 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.656815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.656824 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.656843 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.656855 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.662853 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfgfr\" (UniqueName: \"kubernetes.io/projected/9e62a642-7ddb-45eb-ba9a-efb54a4f14aa-kube-api-access-nfgfr\") pod \"ovnkube-control-plane-749d76644c-qq4js\" (UID: \"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.693966 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=3.693939439 podStartE2EDuration="3.693939439s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:04.638880469 +0000 UTC m=+21.266346797" watchObservedRunningTime="2026-01-06 14:37:04.693939439 +0000 UTC m=+21.321405797" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.710246 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:04 crc kubenswrapper[4744]: E0106 14:37:04.710358 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.710246 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:04 crc kubenswrapper[4744]: E0106 14:37:04.710539 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.733008 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.733054 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm8fs\" (UniqueName: \"kubernetes.io/projected/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-kube-api-access-gm8fs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:04 crc kubenswrapper[4744]: E0106 14:37:04.733264 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:04 crc kubenswrapper[4744]: E0106 14:37:04.733377 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs podName:4e45835e-b6fd-4d25-ad1a-c5d9b9221cde nodeName:}" failed. No retries permitted until 2026-01-06 14:37:05.233349828 +0000 UTC m=+21.860816146 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs") pod "network-metrics-daemon-f9m2v" (UID: "4e45835e-b6fd-4d25-ad1a-c5d9b9221cde") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.759686 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.759753 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.759778 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.759807 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.759822 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.763226 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm8fs\" (UniqueName: \"kubernetes.io/projected/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-kube-api-access-gm8fs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.765764 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" Jan 06 14:37:04 crc kubenswrapper[4744]: W0106 14:37:04.784354 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e62a642_7ddb_45eb_ba9a_efb54a4f14aa.slice/crio-5ee3144ee659597fe3d1f8a6c117b97839437f54c4a7f7cfd26620e23b52cc00 WatchSource:0}: Error finding container 5ee3144ee659597fe3d1f8a6c117b97839437f54c4a7f7cfd26620e23b52cc00: Status 404 returned error can't find the container with id 5ee3144ee659597fe3d1f8a6c117b97839437f54c4a7f7cfd26620e23b52cc00 Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.851211 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.85112697 podStartE2EDuration="3.85112697s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:04.85074924 +0000 UTC m=+21.478215588" watchObservedRunningTime="2026-01-06 14:37:04.85112697 +0000 UTC m=+21.478593338" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.863351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.863395 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.863404 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.863416 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.863425 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.939690 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" event={"ID":"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa","Type":"ContainerStarted","Data":"5ee3144ee659597fe3d1f8a6c117b97839437f54c4a7f7cfd26620e23b52cc00"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.943932 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerStarted","Data":"dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.943999 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerStarted","Data":"30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.944010 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerStarted","Data":"4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.945444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-8mz44" event={"ID":"38f0ac6a-e5ca-48e4-960f-371ac91370a6","Type":"ContainerStarted","Data":"952b1dcce54141e17c8220e8dbe0a0ae7d59954c0dfbb09130bc157173a73e9a"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.946920 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-98zj4" event={"ID":"9fc78f49-6e10-4a01-a37b-8de049da84ae","Type":"ContainerStarted","Data":"381293142f4f8f17f15308deb777a62ece6ffcc76fb62de0e35bcf6f6ed255a6"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.946981 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-98zj4" event={"ID":"9fc78f49-6e10-4a01-a37b-8de049da84ae","Type":"ContainerStarted","Data":"611f75dedf2a5fdc1b13a5ff2c4c2bc5eb90fd2ede8fba702935d11bd703a307"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.948761 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4e53714-2c82-4edb-aafc-cc9f3890a645" containerID="18011149894ad508f912b4bd206a34496ed3afb612846f94985c7984ab6c793d" exitCode=0 Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.948883 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" event={"ID":"e4e53714-2c82-4edb-aafc-cc9f3890a645","Type":"ContainerDied","Data":"18011149894ad508f912b4bd206a34496ed3afb612846f94985c7984ab6c793d"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.961209 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-8mz44" podStartSLOduration=3.961196219 podStartE2EDuration="3.961196219s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:04.961005804 +0000 UTC m=+21.588472202" watchObservedRunningTime="2026-01-06 14:37:04.961196219 +0000 UTC m=+21.588662537" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.967755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.967806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.967815 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.967833 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.967845 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:04Z","lastTransitionTime":"2026-01-06T14:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:04 crc kubenswrapper[4744]: I0106 14:37:04.996361 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-98zj4" podStartSLOduration=3.996331464 podStartE2EDuration="3.996331464s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:04.995290817 +0000 UTC m=+21.622757135" watchObservedRunningTime="2026-01-06 14:37:04.996331464 +0000 UTC m=+21.623797782" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.071554 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.071606 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.071619 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.071641 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.071653 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:05Z","lastTransitionTime":"2026-01-06T14:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.174291 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.174337 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.174349 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.174372 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.174383 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:05Z","lastTransitionTime":"2026-01-06T14:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.239667 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.239861 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.239925 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs podName:4e45835e-b6fd-4d25-ad1a-c5d9b9221cde nodeName:}" failed. No retries permitted until 2026-01-06 14:37:06.23991179 +0000 UTC m=+22.867378108 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs") pod "network-metrics-daemon-f9m2v" (UID: "4e45835e-b6fd-4d25-ad1a-c5d9b9221cde") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.281368 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.281418 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.281430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.281453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.281470 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:05Z","lastTransitionTime":"2026-01-06T14:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.340493 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.340855 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:09.340829308 +0000 UTC m=+25.968295626 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.384120 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.384216 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.384235 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.384263 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.384280 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:05Z","lastTransitionTime":"2026-01-06T14:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.442273 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.442341 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.442376 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.442404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442475 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442496 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442510 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442524 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442548 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442568 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:09.442553007 +0000 UTC m=+26.070019325 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442584 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:09.442578098 +0000 UTC m=+26.070044406 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442610 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:09.442593218 +0000 UTC m=+26.070059546 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442693 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442752 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442772 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.442879 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:09.442846875 +0000 UTC m=+26.070313403 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.487793 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.487853 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.487866 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.487889 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.487901 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:05Z","lastTransitionTime":"2026-01-06T14:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.591094 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.591153 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.591184 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.591213 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.591228 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:05Z","lastTransitionTime":"2026-01-06T14:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.657747 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.663937 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.672614 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.694353 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.694394 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.694405 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.694422 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.694435 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:05Z","lastTransitionTime":"2026-01-06T14:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.710951 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.710963 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.711110 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.711356 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f9m2v" podUID="4e45835e-b6fd-4d25-ad1a-c5d9b9221cde" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.796874 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.796922 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.796937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.796954 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.796966 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:05Z","lastTransitionTime":"2026-01-06T14:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.900186 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.900239 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.900260 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.900284 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.900302 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:05Z","lastTransitionTime":"2026-01-06T14:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.922465 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.969265 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4e53714-2c82-4edb-aafc-cc9f3890a645" containerID="eddf1e421abd2c2e52abcdea9382ead42ef4df55997a595af68ebaa0821b943f" exitCode=0 Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.969337 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" event={"ID":"e4e53714-2c82-4edb-aafc-cc9f3890a645","Type":"ContainerDied","Data":"eddf1e421abd2c2e52abcdea9382ead42ef4df55997a595af68ebaa0821b943f"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.973050 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" event={"ID":"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa","Type":"ContainerStarted","Data":"96851e554eafe20da8f3411f35a3f0d4b09baf2fc5fb12175506ef2d19c779f0"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.973253 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" event={"ID":"9e62a642-7ddb-45eb-ba9a-efb54a4f14aa","Type":"ContainerStarted","Data":"bca225f9f560abb7f3fbc3aaac5d7413ed4eeb8578f66151ced1e51cdc591a9a"} Jan 06 14:37:05 crc kubenswrapper[4744]: I0106 14:37:05.975585 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"dceb9a52590a93dc5481da24f0681ee104d81afe9330bdcf379d739daa560820"} Jan 06 14:37:05 crc kubenswrapper[4744]: E0106 14:37:05.987832 4744 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.003677 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.004232 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.004255 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.004281 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.004301 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.049030 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=1.04898379 podStartE2EDuration="1.04898379s" podCreationTimestamp="2026-01-06 14:37:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:06.01026713 +0000 UTC m=+22.637733478" watchObservedRunningTime="2026-01-06 14:37:06.04898379 +0000 UTC m=+22.676450118" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.063634 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-qq4js" podStartSLOduration=4.063610365 podStartE2EDuration="4.063610365s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:06.063256936 +0000 UTC m=+22.690723274" watchObservedRunningTime="2026-01-06 14:37:06.063610365 +0000 UTC m=+22.691076693" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.108846 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.108901 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.108914 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.108932 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.108945 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.211968 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.212028 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.212045 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.212068 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.212085 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.251206 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:06 crc kubenswrapper[4744]: E0106 14:37:06.251366 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:06 crc kubenswrapper[4744]: E0106 14:37:06.251417 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs podName:4e45835e-b6fd-4d25-ad1a-c5d9b9221cde nodeName:}" failed. No retries permitted until 2026-01-06 14:37:08.25140288 +0000 UTC m=+24.878869198 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs") pod "network-metrics-daemon-f9m2v" (UID: "4e45835e-b6fd-4d25-ad1a-c5d9b9221cde") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.314522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.314589 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.314599 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.314618 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.314629 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.360869 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.405031 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.417296 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.417378 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.417402 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.417433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.417457 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.521114 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.521455 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.521557 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.521658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.521748 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.624971 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.625314 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.625595 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.625691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.625985 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.710738 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.710751 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:06 crc kubenswrapper[4744]: E0106 14:37:06.710902 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 06 14:37:06 crc kubenswrapper[4744]: E0106 14:37:06.711080 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.729935 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.730355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.730524 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.730659 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.730771 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.834670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.834716 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.834726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.834742 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.834756 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.936965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.937351 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.937498 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.937654 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.937810 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:06Z","lastTransitionTime":"2026-01-06T14:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.985071 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4e53714-2c82-4edb-aafc-cc9f3890a645" containerID="7714567aeceed8ba043f9844aa1b3f3abea7b1a3e7629130b525eb4e2c3aa7b4" exitCode=0 Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.985225 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" event={"ID":"e4e53714-2c82-4edb-aafc-cc9f3890a645","Type":"ContainerDied","Data":"7714567aeceed8ba043f9844aa1b3f3abea7b1a3e7629130b525eb4e2c3aa7b4"} Jan 06 14:37:06 crc kubenswrapper[4744]: I0106 14:37:06.992953 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerStarted","Data":"0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.041529 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.041608 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.041628 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.041662 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.041683 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.145453 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.145507 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.145519 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.145540 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.145555 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.250997 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.251058 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.251078 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.251132 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.251152 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.354384 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.354422 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.354430 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.354447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.354458 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.457228 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.457273 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.457282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.457297 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.457308 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.560937 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.560986 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.561004 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.561027 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.561045 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.664044 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.664110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.664128 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.664149 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.664194 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.710737 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:07 crc kubenswrapper[4744]: E0106 14:37:07.710911 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.710989 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:07 crc kubenswrapper[4744]: E0106 14:37:07.711262 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f9m2v" podUID="4e45835e-b6fd-4d25-ad1a-c5d9b9221cde" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.767071 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.767522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.767706 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.767863 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.768016 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.871795 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.871837 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.871847 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.871865 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.871875 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.974020 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.974079 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.974093 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.974110 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:07 crc kubenswrapper[4744]: I0106 14:37:07.974123 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:07Z","lastTransitionTime":"2026-01-06T14:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.002504 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4e53714-2c82-4edb-aafc-cc9f3890a645" containerID="8429bbaf21ac9dd1b4abd231757994ad1a23cbd4bdac97b818cee827bde0df3c" exitCode=0 Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.002587 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" event={"ID":"e4e53714-2c82-4edb-aafc-cc9f3890a645","Type":"ContainerDied","Data":"8429bbaf21ac9dd1b4abd231757994ad1a23cbd4bdac97b818cee827bde0df3c"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.077615 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.077658 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.077668 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.077684 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.077696 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:08Z","lastTransitionTime":"2026-01-06T14:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.180884 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.180962 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.180982 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.180999 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.181009 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:08Z","lastTransitionTime":"2026-01-06T14:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.277458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:08 crc kubenswrapper[4744]: E0106 14:37:08.277941 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:08 crc kubenswrapper[4744]: E0106 14:37:08.278073 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs podName:4e45835e-b6fd-4d25-ad1a-c5d9b9221cde nodeName:}" failed. No retries permitted until 2026-01-06 14:37:12.27804382 +0000 UTC m=+28.905510178 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs") pod "network-metrics-daemon-f9m2v" (UID: "4e45835e-b6fd-4d25-ad1a-c5d9b9221cde") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.284433 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.284469 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.284481 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.284503 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.284514 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:08Z","lastTransitionTime":"2026-01-06T14:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.387659 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.387726 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.387743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.387800 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.387818 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:08Z","lastTransitionTime":"2026-01-06T14:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.491688 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.491755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.491772 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.491799 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.491817 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:08Z","lastTransitionTime":"2026-01-06T14:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.595076 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.595477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.595605 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.595766 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.595890 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:08Z","lastTransitionTime":"2026-01-06T14:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.699098 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.699145 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.699191 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.699209 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.699222 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:08Z","lastTransitionTime":"2026-01-06T14:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.710549 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:08 crc kubenswrapper[4744]: E0106 14:37:08.710987 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.711351 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:08 crc kubenswrapper[4744]: E0106 14:37:08.711472 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.802426 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.802923 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.802941 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.802965 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.802985 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:08Z","lastTransitionTime":"2026-01-06T14:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.906695 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.906921 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.907058 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.907233 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:08 crc kubenswrapper[4744]: I0106 14:37:08.907392 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:08Z","lastTransitionTime":"2026-01-06T14:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.009752 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.009813 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.009831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.009861 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.009884 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.013292 4744 generic.go:334] "Generic (PLEG): container finished" podID="e4e53714-2c82-4edb-aafc-cc9f3890a645" containerID="948a67199379baff67894b3dd65b6833bf83b021895a9e20aa7bb1cda39d2b5f" exitCode=0 Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.013355 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" event={"ID":"e4e53714-2c82-4edb-aafc-cc9f3890a645","Type":"ContainerDied","Data":"948a67199379baff67894b3dd65b6833bf83b021895a9e20aa7bb1cda39d2b5f"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.112681 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.112768 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.112800 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.112829 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.112848 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.224792 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.224831 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.224841 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.224859 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.224872 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.328447 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.328510 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.328534 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.328562 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.328580 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.391351 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.391697 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:17.39162662 +0000 UTC m=+34.019092968 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.433392 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.433460 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.433477 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.433505 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.433541 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.492924 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.493236 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.493280 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.493335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493447 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493504 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493527 4744 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493570 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493569 4744 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493643 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:17.493609596 +0000 UTC m=+34.121075944 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493688 4744 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493739 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:17.493694948 +0000 UTC m=+34.121161476 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493601 4744 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493783 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:17.49376373 +0000 UTC m=+34.121230338 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493790 4744 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.493868 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:17.493838972 +0000 UTC m=+34.121305490 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.537091 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.537151 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.537202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.537229 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.537248 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.640397 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.640472 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.640485 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.640550 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.640565 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.710508 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.710717 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.710819 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:09 crc kubenswrapper[4744]: E0106 14:37:09.711066 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f9m2v" podUID="4e45835e-b6fd-4d25-ad1a-c5d9b9221cde" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.743596 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.743670 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.743693 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.743719 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.743744 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.847202 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.847280 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.847298 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.847330 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.847352 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.950727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.950806 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.950830 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.950860 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:09 crc kubenswrapper[4744]: I0106 14:37:09.950879 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:09Z","lastTransitionTime":"2026-01-06T14:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.025822 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerStarted","Data":"b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.026252 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.033300 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" event={"ID":"e4e53714-2c82-4edb-aafc-cc9f3890a645","Type":"ContainerStarted","Data":"870e28bcaa9d2fe96eba759873719dc047d6d519fa1913df90e36dce8e2ade1d"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.102319 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.102729 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.102747 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.102770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.102788 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:10Z","lastTransitionTime":"2026-01-06T14:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.106646 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.158628 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" podStartSLOduration=9.15860726 podStartE2EDuration="9.15860726s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:10.114238902 +0000 UTC m=+26.741705240" watchObservedRunningTime="2026-01-06 14:37:10.15860726 +0000 UTC m=+26.786073588" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.186660 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-jsxwz" podStartSLOduration=9.186638989 podStartE2EDuration="9.186638989s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:10.186002012 +0000 UTC m=+26.813468370" watchObservedRunningTime="2026-01-06 14:37:10.186638989 +0000 UTC m=+26.814105327" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.205294 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.205352 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.205365 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.205383 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.205396 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:10Z","lastTransitionTime":"2026-01-06T14:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.310262 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.310313 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.310330 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.310354 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.310373 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:10Z","lastTransitionTime":"2026-01-06T14:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.412461 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.412522 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.412538 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.412569 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.412592 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:10Z","lastTransitionTime":"2026-01-06T14:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.515610 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.515690 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.515714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.515743 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.515763 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:10Z","lastTransitionTime":"2026-01-06T14:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.618598 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.618666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.618691 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.618718 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.618742 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:10Z","lastTransitionTime":"2026-01-06T14:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.710109 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.710143 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:10 crc kubenswrapper[4744]: E0106 14:37:10.710366 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 06 14:37:10 crc kubenswrapper[4744]: E0106 14:37:10.710546 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.721282 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.721338 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.721355 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.721379 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.721395 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:10Z","lastTransitionTime":"2026-01-06T14:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.824667 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.824736 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.824755 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.824781 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.824801 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:10Z","lastTransitionTime":"2026-01-06T14:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.927489 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.927544 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.927556 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.927574 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:10 crc kubenswrapper[4744]: I0106 14:37:10.927586 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:10Z","lastTransitionTime":"2026-01-06T14:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.030930 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.030980 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.030992 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.031010 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.031032 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:11Z","lastTransitionTime":"2026-01-06T14:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.035877 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.036572 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.069070 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.134358 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.134432 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.134450 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.134476 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.134497 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:11Z","lastTransitionTime":"2026-01-06T14:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.237770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.237850 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.237876 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.237910 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.237934 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:11Z","lastTransitionTime":"2026-01-06T14:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.341666 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.341727 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.341745 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.341770 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.341794 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:11Z","lastTransitionTime":"2026-01-06T14:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.352667 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-f9m2v"] Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.352838 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:11 crc kubenswrapper[4744]: E0106 14:37:11.352977 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f9m2v" podUID="4e45835e-b6fd-4d25-ad1a-c5d9b9221cde" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.445005 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.445051 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.445063 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.445079 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.445094 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:11Z","lastTransitionTime":"2026-01-06T14:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.547559 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.547651 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.547680 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.547714 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.547738 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:11Z","lastTransitionTime":"2026-01-06T14:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.650991 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.651064 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.651082 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.651112 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.651130 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:11Z","lastTransitionTime":"2026-01-06T14:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.711788 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:11 crc kubenswrapper[4744]: E0106 14:37:11.711955 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.754999 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.755088 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.755117 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.755152 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.755213 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:11Z","lastTransitionTime":"2026-01-06T14:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.782130 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.782227 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.782253 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.782285 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.782305 4744 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-06T14:37:11Z","lastTransitionTime":"2026-01-06T14:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.887927 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk"] Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.888507 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.890826 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.891113 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.893032 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 06 14:37:11 crc kubenswrapper[4744]: I0106 14:37:11.893145 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.022374 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.022420 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.022476 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.022499 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.022531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.038832 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.123737 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.123810 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.123865 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.123949 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.123999 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.124140 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.124151 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.125272 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.145796 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.163307 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3c093c06-0e18-4bbe-b0c2-6b91df4206c8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fh8lk\" (UID: \"3c093c06-0e18-4bbe-b0c2-6b91df4206c8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.206623 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.325876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:12 crc kubenswrapper[4744]: E0106 14:37:12.326504 4744 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:12 crc kubenswrapper[4744]: E0106 14:37:12.326666 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs podName:4e45835e-b6fd-4d25-ad1a-c5d9b9221cde nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.326626643 +0000 UTC m=+36.954092991 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs") pod "network-metrics-daemon-f9m2v" (UID: "4e45835e-b6fd-4d25-ad1a-c5d9b9221cde") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.710738 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:12 crc kubenswrapper[4744]: E0106 14:37:12.710846 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 06 14:37:12 crc kubenswrapper[4744]: I0106 14:37:12.711006 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:12 crc kubenswrapper[4744]: E0106 14:37:12.711197 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 06 14:37:13 crc kubenswrapper[4744]: I0106 14:37:13.043104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" event={"ID":"3c093c06-0e18-4bbe-b0c2-6b91df4206c8","Type":"ContainerStarted","Data":"29e57d9dc7e75b10c9967701b9a1ea4ee766946c841cfe90445b1f086e5a92bc"} Jan 06 14:37:13 crc kubenswrapper[4744]: I0106 14:37:13.043198 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" event={"ID":"3c093c06-0e18-4bbe-b0c2-6b91df4206c8","Type":"ContainerStarted","Data":"1ccb67db38848b3496be7480d872d9bc335978d20fe151006c62d8991de79de7"} Jan 06 14:37:13 crc kubenswrapper[4744]: I0106 14:37:13.043257 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 06 14:37:13 crc kubenswrapper[4744]: I0106 14:37:13.069236 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fh8lk" podStartSLOduration=12.069217022 podStartE2EDuration="12.069217022s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:13.066194873 +0000 UTC m=+29.693661221" watchObservedRunningTime="2026-01-06 14:37:13.069217022 +0000 UTC m=+29.696683380" Jan 06 14:37:13 crc kubenswrapper[4744]: I0106 14:37:13.711008 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:13 crc kubenswrapper[4744]: I0106 14:37:13.711146 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:13 crc kubenswrapper[4744]: E0106 14:37:13.712910 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-f9m2v" podUID="4e45835e-b6fd-4d25-ad1a-c5d9b9221cde" Jan 06 14:37:13 crc kubenswrapper[4744]: E0106 14:37:13.713041 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.541994 4744 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.542454 4744 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.603118 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.603802 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.605027 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.606009 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.606439 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9z7kv"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.607318 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.608444 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.609004 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.610701 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.611277 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.611628 4744 reflector.go:561] object-"openshift-route-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.611688 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.611905 4744 reflector.go:561] object-"openshift-route-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.612025 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.612242 4744 reflector.go:561] object-"openshift-route-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.612284 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.612565 4744 reflector.go:561] object-"openshift-route-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.612672 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.612803 4744 reflector.go:561] object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2": failed to list *v1.Secret: secrets "route-controller-manager-sa-dockercfg-h2zr2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.612894 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-h2zr2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"route-controller-manager-sa-dockercfg-h2zr2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.615743 4744 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template": failed to list *v1.Secret: secrets "v4-0-config-system-ocp-branding-template" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.615791 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-ocp-branding-template\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-ocp-branding-template\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.615853 4744 reflector.go:561] object-"openshift-authentication"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.615872 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.615924 4744 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-error": failed to list *v1.Secret: secrets "v4-0-config-user-template-error" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.615943 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-error\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.616014 4744 reflector.go:561] object-"openshift-oauth-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.616033 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.616081 4744 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.616099 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.616226 4744 reflector.go:561] object-"openshift-oauth-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.616248 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.616298 4744 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-cliconfig": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-cliconfig" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.616316 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-cliconfig\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.621553 4744 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.621639 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.621853 4744 reflector.go:561] object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq": failed to list *v1.Secret: secrets "oauth-apiserver-sa-dockercfg-6r2bq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.621880 4744 reflector.go:561] object-"openshift-route-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.622100 4744 reflector.go:561] object-"openshift-authentication"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.622216 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.621936 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-6r2bq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-apiserver-sa-dockercfg-6r2bq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.622234 4744 reflector.go:561] object-"openshift-oauth-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.622448 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.622493 4744 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-login": failed to list *v1.Secret: secrets "v4-0-config-user-template-login" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.622599 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-login\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.622078 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.622236 4744 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.622819 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.622299 4744 reflector.go:561] object-"openshift-oauth-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.622977 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.622360 4744 reflector.go:561] object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc": failed to list *v1.Secret: secrets "oauth-openshift-dockercfg-znhcc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.623130 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-znhcc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-openshift-dockercfg-znhcc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.623414 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2hcpt"] Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.623518 4744 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-serving-cert": failed to list *v1.Secret: secrets "v4-0-config-system-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.623595 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.624250 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.625246 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pc7r"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.625922 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.626265 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mpb2d"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.627131 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.628732 4744 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data": failed to list *v1.Secret: secrets "v4-0-config-user-idp-0-file-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.628848 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-idp-0-file-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.629045 4744 reflector.go:561] object-"openshift-oauth-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.629114 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.629568 4744 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-provider-selection": failed to list *v1.Secret: secrets "v4-0-config-user-template-provider-selection" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.629624 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-provider-selection\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.629704 4744 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-service-ca": failed to list *v1.ConfigMap: configmaps "v4-0-config-system-service-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.629739 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"v4-0-config-system-service-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: W0106 14:37:14.629815 4744 reflector.go:561] object-"openshift-authentication"/"audit": failed to list *v1.ConfigMap: configmaps "audit" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Jan 06 14:37:14 crc kubenswrapper[4744]: E0106 14:37:14.629844 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"audit\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.645868 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.646140 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.647617 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.648271 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.652541 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.652791 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.653243 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.653386 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.653987 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.654687 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.656424 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qplq4"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.656729 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.657039 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.660578 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.663348 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.666489 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2qvhp"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.683688 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.684272 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.684346 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.684406 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.684549 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.684871 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.685412 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-bbfrh"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.685534 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.685950 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686077 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686141 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686245 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686292 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-6pgv4"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686335 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686442 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686458 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686646 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686701 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686769 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bbfrh" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686781 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.687048 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.686784 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.687276 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.687312 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.687490 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.687652 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.687659 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.687710 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.687994 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vnmvf"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.688057 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.688670 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.689855 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z84zq"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.690653 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.691038 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.691101 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.691255 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.691652 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-v9kk6"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.692258 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.692730 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.692783 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.693100 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8ltv4"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.693343 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.710758 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.711147 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.711262 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.711287 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.712302 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-tgntc"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.713012 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.714170 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.715230 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.715310 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.715337 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9m5ds"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.715404 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.715525 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.716811 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.717060 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.717261 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.764255 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx77r\" (UniqueName: \"kubernetes.io/projected/c031a005-c650-4129-b764-55673c0e3ddd-kube-api-access-jx77r\") pod \"openshift-apiserver-operator-796bbdcf4f-7d8pp\" (UID: \"c031a005-c650-4129-b764-55673c0e3ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.764301 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.782969 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.783256 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.783343 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.783527 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.783842 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.783958 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.783974 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.784188 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.784277 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.784287 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.784437 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.784539 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.784607 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.784642 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.784879 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.784934 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.785052 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.785413 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.785819 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.786252 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.786460 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.786608 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.787000 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.788850 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.789018 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.789134 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.789201 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.789810 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s5h9w"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.790265 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.790465 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.790750 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.790958 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.791086 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.791230 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.791369 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.791489 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.791628 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.791780 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.791896 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.792017 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.792172 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.792298 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.796327 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.796433 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.796597 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.796735 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.796905 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.797028 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.799406 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.799494 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.799623 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.799823 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.799899 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.800113 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.801966 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.799912 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9z7kv"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.802233 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgpw"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.802775 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.804696 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.804941 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.805843 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.805923 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.806055 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.806326 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.809940 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810352 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810486 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810645 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810695 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810862 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1d1d915-9b82-4311-974c-ea3d5e58bf80-serving-cert\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810904 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c031a005-c650-4129-b764-55673c0e3ddd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7d8pp\" (UID: \"c031a005-c650-4129-b764-55673c0e3ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810922 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810942 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a67f83fd-2c95-427b-afd8-bc19d5abda9e-node-pullsecrets\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810957 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsj5c\" (UniqueName: \"kubernetes.io/projected/8430f458-83cd-4453-bb18-883d49bae7d9-kube-api-access-qsj5c\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.810973 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55gwc\" (UniqueName: \"kubernetes.io/projected/f1d1d915-9b82-4311-974c-ea3d5e58bf80-kube-api-access-55gwc\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811029 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811066 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5e458d23-aef6-4e91-9741-46d74c403131-images\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811105 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt2l9\" (UniqueName: \"kubernetes.io/projected/a67f83fd-2c95-427b-afd8-bc19d5abda9e-kube-api-access-zt2l9\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811121 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e458d23-aef6-4e91-9741-46d74c403131-config\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811139 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8430f458-83cd-4453-bb18-883d49bae7d9-config\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811190 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-config\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811264 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811279 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-audit-dir\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811297 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811316 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-config\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811331 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-image-import-ca\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811347 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/87c264b6-7ed1-4b76-bc4b-a2c4e61ac543-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k4q45\" (UID: \"87c264b6-7ed1-4b76-bc4b-a2c4e61ac543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811367 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b683331c-2c30-482e-a85d-975669ae737e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zk6z4\" (UID: \"b683331c-2c30-482e-a85d-975669ae737e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811381 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811539 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811556 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811576 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811594 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4vhc\" (UniqueName: \"kubernetes.io/projected/5402a80f-2077-4207-81bc-cc12a2935886-kube-api-access-p4vhc\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811610 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811624 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a67f83fd-2c95-427b-afd8-bc19d5abda9e-encryption-config\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811639 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv8qd\" (UniqueName: \"kubernetes.io/projected/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-kube-api-access-vv8qd\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-etcd-serving-ca\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811749 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26427892-f331-42fd-98a5-687d7efe8d6a-config\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811773 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811796 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-client-ca\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811839 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811862 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811881 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-encryption-config\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811917 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26427892-f331-42fd-98a5-687d7efe8d6a-trusted-ca\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8430f458-83cd-4453-bb18-883d49bae7d9-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811951 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-dir\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.811984 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26427892-f331-42fd-98a5-687d7efe8d6a-serving-cert\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812007 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f8df\" (UniqueName: \"kubernetes.io/projected/26427892-f331-42fd-98a5-687d7efe8d6a-kube-api-access-9f8df\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812021 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-client\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812041 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25dcd\" (UniqueName: \"kubernetes.io/projected/5e458d23-aef6-4e91-9741-46d74c403131-kube-api-access-25dcd\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812065 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b683331c-2c30-482e-a85d-975669ae737e-proxy-tls\") pod \"machine-config-controller-84d6567774-zk6z4\" (UID: \"b683331c-2c30-482e-a85d-975669ae737e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812082 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812098 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a67f83fd-2c95-427b-afd8-bc19d5abda9e-etcd-client\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812254 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2nh4\" (UniqueName: \"kubernetes.io/projected/87c264b6-7ed1-4b76-bc4b-a2c4e61ac543-kube-api-access-s2nh4\") pod \"cluster-samples-operator-665b6dd947-k4q45\" (UID: \"87c264b6-7ed1-4b76-bc4b-a2c4e61ac543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812275 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a67f83fd-2c95-427b-afd8-bc19d5abda9e-serving-cert\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812292 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8430f458-83cd-4453-bb18-883d49bae7d9-service-ca-bundle\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812310 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812342 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5e458d23-aef6-4e91-9741-46d74c403131-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812367 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c031a005-c650-4129-b764-55673c0e3ddd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7d8pp\" (UID: \"c031a005-c650-4129-b764-55673c0e3ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812443 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gznpj\" (UniqueName: \"kubernetes.io/projected/82c5f022-eee7-42a1-b9b3-e00c07783034-kube-api-access-gznpj\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812461 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-serving-cert\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812484 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812509 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812678 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-audit\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812771 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-client-ca\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.812895 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8430f458-83cd-4453-bb18-883d49bae7d9-serving-cert\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.813026 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a67f83fd-2c95-427b-afd8-bc19d5abda9e-audit-dir\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.813139 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-audit-policies\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.813457 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k6g9\" (UniqueName: \"kubernetes.io/projected/b683331c-2c30-482e-a85d-975669ae737e-kube-api-access-4k6g9\") pod \"machine-config-controller-84d6567774-zk6z4\" (UID: \"b683331c-2c30-482e-a85d-975669ae737e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.813653 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2hcpt"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.813727 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.834330 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.837242 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.837440 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.839865 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.843830 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.846868 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.862938 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-6xk62"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.864235 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.864405 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bbfrh"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.866860 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z84zq"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.867881 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2qvhp"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.868996 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mpb2d"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.869646 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.869833 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.873714 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vnmvf"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.874415 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9l4fk"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.875726 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8ltv4"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.875823 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.876299 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.877323 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.879424 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.880180 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6pgv4"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.884936 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qplq4"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.886403 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.887282 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.887380 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-bd9mm"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.888068 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.890230 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-52sxd"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.891288 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.891513 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s5h9w"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.893489 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.894991 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-8wjqf"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.895748 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8wjqf" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.897627 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.898829 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pc7r"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.900051 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9m5ds"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.901367 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.902897 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.904255 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.905487 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-v9kk6"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.905868 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.906622 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8wjqf"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.907810 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.908923 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.911126 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgpw"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.911410 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.912736 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914053 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914358 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914409 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-console-config\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914438 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1ca29946-9c57-462b-b53e-1bb596bfa3bb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9m5ds\" (UID: \"1ca29946-9c57-462b-b53e-1bb596bfa3bb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914468 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsj5c\" (UniqueName: \"kubernetes.io/projected/8430f458-83cd-4453-bb18-883d49bae7d9-kube-api-access-qsj5c\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914493 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55gwc\" (UniqueName: \"kubernetes.io/projected/f1d1d915-9b82-4311-974c-ea3d5e58bf80-kube-api-access-55gwc\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914563 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-audit-dir\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914591 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914616 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914641 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-audit-dir\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/40203b7c-e649-4742-b0b4-6963fb2e1aca-default-certificate\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914712 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e615c362-aa4f-488f-ba46-9fbb9c66e204-profile-collector-cert\") pod \"catalog-operator-68c6474976-6fnxs\" (UID: \"e615c362-aa4f-488f-ba46-9fbb9c66e204\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914761 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlrh9\" (UniqueName: \"kubernetes.io/projected/1ca29946-9c57-462b-b53e-1bb596bfa3bb-kube-api-access-tlrh9\") pod \"multus-admission-controller-857f4d67dd-9m5ds\" (UID: \"1ca29946-9c57-462b-b53e-1bb596bfa3bb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914795 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d42baab-089b-497d-8a9f-aea1966a66eb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sp97v\" (UID: \"4d42baab-089b-497d-8a9f-aea1966a66eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914852 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/87c264b6-7ed1-4b76-bc4b-a2c4e61ac543-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k4q45\" (UID: \"87c264b6-7ed1-4b76-bc4b-a2c4e61ac543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914890 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-trusted-ca-bundle\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914933 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn9xj\" (UniqueName: \"kubernetes.io/projected/ed5d8f62-5fc4-46f9-bc8b-0154549a9222-kube-api-access-kn9xj\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdt2p\" (UID: \"ed5d8f62-5fc4-46f9-bc8b-0154549a9222\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.914972 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96qqt\" (UniqueName: \"kubernetes.io/projected/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-kube-api-access-96qqt\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.915042 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.915132 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.915369 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.915535 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.915713 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.915805 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.916028 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt2tn\" (UniqueName: \"kubernetes.io/projected/e5fc7754-5768-4bda-9534-15bd44890e2c-kube-api-access-gt2tn\") pod \"openshift-config-operator-7777fb866f-z84zq\" (UID: \"e5fc7754-5768-4bda-9534-15bd44890e2c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.916171 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhtfp\" (UniqueName: \"kubernetes.io/projected/4d42baab-089b-497d-8a9f-aea1966a66eb-kube-api-access-mhtfp\") pod \"openshift-controller-manager-operator-756b6f6bc6-sp97v\" (UID: \"4d42baab-089b-497d-8a9f-aea1966a66eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.916298 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26427892-f331-42fd-98a5-687d7efe8d6a-trusted-ca\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.917641 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-encryption-config\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.917736 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26427892-f331-42fd-98a5-687d7efe8d6a-serving-cert\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.917807 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-dir\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.917886 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25dcd\" (UniqueName: \"kubernetes.io/projected/5e458d23-aef6-4e91-9741-46d74c403131-kube-api-access-25dcd\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.917970 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b683331c-2c30-482e-a85d-975669ae737e-proxy-tls\") pod \"machine-config-controller-84d6567774-zk6z4\" (UID: \"b683331c-2c30-482e-a85d-975669ae737e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.918052 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.918189 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rv5c\" (UniqueName: \"kubernetes.io/projected/e615c362-aa4f-488f-ba46-9fbb9c66e204-kube-api-access-4rv5c\") pod \"catalog-operator-68c6474976-6fnxs\" (UID: \"e615c362-aa4f-488f-ba46-9fbb9c66e204\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.918270 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-etcd-service-ca\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.918376 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c031a005-c650-4129-b764-55673c0e3ddd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7d8pp\" (UID: \"c031a005-c650-4129-b764-55673c0e3ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.918462 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-etcd-ca\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.917519 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.916916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.917597 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26427892-f331-42fd-98a5-687d7efe8d6a-trusted-ca\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.917843 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-dir\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.918766 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae0463e-7748-4f59-8074-91cd1364115f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rhjr7\" (UID: \"4ae0463e-7748-4f59-8074-91cd1364115f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919173 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4c680757-4206-4c9f-ba2b-d5443aae9f94-metrics-tls\") pod \"dns-operator-744455d44c-v9kk6\" (UID: \"4c680757-4206-4c9f-ba2b-d5443aae9f94\") " pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919256 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gznpj\" (UniqueName: \"kubernetes.io/projected/82c5f022-eee7-42a1-b9b3-e00c07783034-kube-api-access-gznpj\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-serving-cert\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-secret-volume\") pod \"collect-profiles-29461830-7v4gm\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919483 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-audit\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919557 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvz77\" (UniqueName: \"kubernetes.io/projected/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-kube-api-access-dvz77\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919646 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-config-volume\") pod \"collect-profiles-29461830-7v4gm\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919724 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pbv8\" (UniqueName: \"kubernetes.io/projected/0640a05a-4a45-4622-8749-88a0621fc348-kube-api-access-5pbv8\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919798 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6a26250c-db32-4acf-aa49-3a6c0d8bf017-metrics-tls\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919874 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a67f83fd-2c95-427b-afd8-bc19d5abda9e-audit-dir\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.919947 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k6g9\" (UniqueName: \"kubernetes.io/projected/b683331c-2c30-482e-a85d-975669ae737e-kube-api-access-4k6g9\") pod \"machine-config-controller-84d6567774-zk6z4\" (UID: \"b683331c-2c30-482e-a85d-975669ae737e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920018 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-oauth-config\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx77r\" (UniqueName: \"kubernetes.io/projected/c031a005-c650-4129-b764-55673c0e3ddd-kube-api-access-jx77r\") pod \"openshift-apiserver-operator-796bbdcf4f-7d8pp\" (UID: \"c031a005-c650-4129-b764-55673c0e3ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920209 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920150 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-audit\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1d1d915-9b82-4311-974c-ea3d5e58bf80-serving-cert\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920458 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-oauth-serving-cert\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920534 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-serving-cert\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920252 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a67f83fd-2c95-427b-afd8-bc19d5abda9e-audit-dir\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920658 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-config\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920736 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c031a005-c650-4129-b764-55673c0e3ddd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7d8pp\" (UID: \"c031a005-c650-4129-b764-55673c0e3ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920806 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-serving-cert\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920884 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gms62\" (UniqueName: \"kubernetes.io/projected/06761c5c-371b-4593-9251-7e374633468c-kube-api-access-gms62\") pod \"migrator-59844c95c7-trj42\" (UID: \"06761c5c-371b-4593-9251-7e374633468c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920956 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5sb8\" (UniqueName: \"kubernetes.io/projected/40203b7c-e649-4742-b0b4-6963fb2e1aca-kube-api-access-g5sb8\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a67f83fd-2c95-427b-afd8-bc19d5abda9e-node-pullsecrets\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921102 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921211 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m8rf\" (UniqueName: \"kubernetes.io/projected/4ae0463e-7748-4f59-8074-91cd1364115f-kube-api-access-8m8rf\") pod \"kube-storage-version-migrator-operator-b67b599dd-rhjr7\" (UID: \"4ae0463e-7748-4f59-8074-91cd1364115f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5e458d23-aef6-4e91-9741-46d74c403131-images\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt2l9\" (UniqueName: \"kubernetes.io/projected/a67f83fd-2c95-427b-afd8-bc19d5abda9e-kube-api-access-zt2l9\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921542 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921308 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/87c264b6-7ed1-4b76-bc4b-a2c4e61ac543-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k4q45\" (UID: \"87c264b6-7ed1-4b76-bc4b-a2c4e61ac543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921137 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a67f83fd-2c95-427b-afd8-bc19d5abda9e-node-pullsecrets\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920994 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c031a005-c650-4129-b764-55673c0e3ddd-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7d8pp\" (UID: \"c031a005-c650-4129-b764-55673c0e3ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.920963 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b683331c-2c30-482e-a85d-975669ae737e-proxy-tls\") pod \"machine-config-controller-84d6567774-zk6z4\" (UID: \"b683331c-2c30-482e-a85d-975669ae737e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921631 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a26250c-db32-4acf-aa49-3a6c0d8bf017-trusted-ca\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.921888 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d42baab-089b-497d-8a9f-aea1966a66eb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sp97v\" (UID: \"4d42baab-089b-497d-8a9f-aea1966a66eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.922030 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e458d23-aef6-4e91-9741-46d74c403131-config\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.923069 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8430f458-83cd-4453-bb18-883d49bae7d9-config\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.923734 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngnnn\" (UniqueName: \"kubernetes.io/projected/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-kube-api-access-ngnnn\") pod \"collect-profiles-29461830-7v4gm\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.923917 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-config\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.924044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/40203b7c-e649-4742-b0b4-6963fb2e1aca-stats-auth\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.924193 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-config\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.922297 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26427892-f331-42fd-98a5-687d7efe8d6a-serving-cert\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.923661 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8430f458-83cd-4453-bb18-883d49bae7d9-config\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.922426 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5e458d23-aef6-4e91-9741-46d74c403131-images\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.933231 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b683331c-2c30-482e-a85d-975669ae737e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zk6z4\" (UID: \"b683331c-2c30-482e-a85d-975669ae737e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934579 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-image-import-ca\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934607 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934634 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a67f83fd-2c95-427b-afd8-bc19d5abda9e-encryption-config\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4vhc\" (UniqueName: \"kubernetes.io/projected/5402a80f-2077-4207-81bc-cc12a2935886-kube-api-access-p4vhc\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e5fc7754-5768-4bda-9534-15bd44890e2c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z84zq\" (UID: \"e5fc7754-5768-4bda-9534-15bd44890e2c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ed5d8f62-5fc4-46f9-bc8b-0154549a9222-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdt2p\" (UID: \"ed5d8f62-5fc4-46f9-bc8b-0154549a9222\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934740 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rtw8\" (UniqueName: \"kubernetes.io/projected/8c3025a2-7183-4772-b64e-97737ccbef30-kube-api-access-7rtw8\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934765 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-etcd-serving-ca\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934792 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26427892-f331-42fd-98a5-687d7efe8d6a-config\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.933245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1d1d915-9b82-4311-974c-ea3d5e58bf80-serving-cert\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934825 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935264 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-client-ca\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv8qd\" (UniqueName: \"kubernetes.io/projected/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-kube-api-access-vv8qd\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935315 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p7jg\" (UniqueName: \"kubernetes.io/projected/4c680757-4206-4c9f-ba2b-d5443aae9f94-kube-api-access-2p7jg\") pod \"dns-operator-744455d44c-v9kk6\" (UID: \"4c680757-4206-4c9f-ba2b-d5443aae9f94\") " pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935383 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6a26250c-db32-4acf-aa49-3a6c0d8bf017-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935434 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935568 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935580 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-config\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935611 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-etcd-client\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ae0463e-7748-4f59-8074-91cd1364115f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rhjr7\" (UID: \"4ae0463e-7748-4f59-8074-91cd1364115f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935688 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f8df\" (UniqueName: \"kubernetes.io/projected/26427892-f331-42fd-98a5-687d7efe8d6a-kube-api-access-9f8df\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935717 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8430f458-83cd-4453-bb18-883d49bae7d9-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935788 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40203b7c-e649-4742-b0b4-6963fb2e1aca-service-ca-bundle\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935815 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c3025a2-7183-4772-b64e-97737ccbef30-config\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935883 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-client\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8c3025a2-7183-4772-b64e-97737ccbef30-machine-approver-tls\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.935967 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a67f83fd-2c95-427b-afd8-bc19d5abda9e-etcd-client\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936018 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-service-ca\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936039 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c3025a2-7183-4772-b64e-97737ccbef30-auth-proxy-config\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936094 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2nh4\" (UniqueName: \"kubernetes.io/projected/87c264b6-7ed1-4b76-bc4b-a2c4e61ac543-kube-api-access-s2nh4\") pod \"cluster-samples-operator-665b6dd947-k4q45\" (UID: \"87c264b6-7ed1-4b76-bc4b-a2c4e61ac543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936174 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e615c362-aa4f-488f-ba46-9fbb9c66e204-srv-cert\") pod \"catalog-operator-68c6474976-6fnxs\" (UID: \"e615c362-aa4f-488f-ba46-9fbb9c66e204\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5e458d23-aef6-4e91-9741-46d74c403131-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a67f83fd-2c95-427b-afd8-bc19d5abda9e-serving-cert\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936363 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26427892-f331-42fd-98a5-687d7efe8d6a-config\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.933296 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936464 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9l4fk"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.933984 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b683331c-2c30-482e-a85d-975669ae737e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zk6z4\" (UID: \"b683331c-2c30-482e-a85d-975669ae737e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.934047 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-config\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936490 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8430f458-83cd-4453-bb18-883d49bae7d9-service-ca-bundle\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936640 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.936654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-image-import-ca\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.923521 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.937424 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-etcd-serving-ca\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.923425 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c031a005-c650-4129-b764-55673c0e3ddd-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7d8pp\" (UID: \"c031a005-c650-4129-b764-55673c0e3ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.922985 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e458d23-aef6-4e91-9741-46d74c403131-config\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.937898 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5fc7754-5768-4bda-9534-15bd44890e2c-serving-cert\") pod \"openshift-config-operator-7777fb866f-z84zq\" (UID: \"e5fc7754-5768-4bda-9534-15bd44890e2c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938107 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-client-ca\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938141 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-client-ca\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938231 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6xk62"] Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938444 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938573 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8430f458-83cd-4453-bb18-883d49bae7d9-serving-cert\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938594 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8430f458-83cd-4453-bb18-883d49bae7d9-service-ca-bundle\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938723 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v68kx\" (UniqueName: \"kubernetes.io/projected/6a26250c-db32-4acf-aa49-3a6c0d8bf017-kube-api-access-v68kx\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938812 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-audit-policies\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938885 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq45t\" (UniqueName: \"kubernetes.io/projected/5f7af32f-7d06-4c00-b613-552287fcb5af-kube-api-access-qq45t\") pod \"downloads-7954f5f757-bbfrh\" (UID: \"5f7af32f-7d06-4c00-b613-552287fcb5af\") " pod="openshift-console/downloads-7954f5f757-bbfrh" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.938978 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40203b7c-e649-4742-b0b4-6963fb2e1aca-metrics-certs\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.939062 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.939429 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-audit-policies\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.939677 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a67f83fd-2c95-427b-afd8-bc19d5abda9e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.939814 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8430f458-83cd-4453-bb18-883d49bae7d9-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.941481 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a67f83fd-2c95-427b-afd8-bc19d5abda9e-encryption-config\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.941522 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.941553 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8430f458-83cd-4453-bb18-883d49bae7d9-serving-cert\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.941562 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a67f83fd-2c95-427b-afd8-bc19d5abda9e-serving-cert\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.941744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/5e458d23-aef6-4e91-9741-46d74c403131-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.943908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a67f83fd-2c95-427b-afd8-bc19d5abda9e-etcd-client\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.946173 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.965617 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 06 14:37:14 crc kubenswrapper[4744]: I0106 14:37:14.986248 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.005903 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.025899 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.040290 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5fc7754-5768-4bda-9534-15bd44890e2c-serving-cert\") pod \"openshift-config-operator-7777fb866f-z84zq\" (UID: \"e5fc7754-5768-4bda-9534-15bd44890e2c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.040368 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v68kx\" (UniqueName: \"kubernetes.io/projected/6a26250c-db32-4acf-aa49-3a6c0d8bf017-kube-api-access-v68kx\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.040394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq45t\" (UniqueName: \"kubernetes.io/projected/5f7af32f-7d06-4c00-b613-552287fcb5af-kube-api-access-qq45t\") pod \"downloads-7954f5f757-bbfrh\" (UID: \"5f7af32f-7d06-4c00-b613-552287fcb5af\") " pod="openshift-console/downloads-7954f5f757-bbfrh" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.040418 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40203b7c-e649-4742-b0b4-6963fb2e1aca-metrics-certs\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.040442 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.040508 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-console-config\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.040532 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1ca29946-9c57-462b-b53e-1bb596bfa3bb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9m5ds\" (UID: \"1ca29946-9c57-462b-b53e-1bb596bfa3bb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.040808 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.041351 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-console-config\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.042411 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.042560 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/40203b7c-e649-4742-b0b4-6963fb2e1aca-default-certificate\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.042812 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e615c362-aa4f-488f-ba46-9fbb9c66e204-profile-collector-cert\") pod \"catalog-operator-68c6474976-6fnxs\" (UID: \"e615c362-aa4f-488f-ba46-9fbb9c66e204\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.042876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlrh9\" (UniqueName: \"kubernetes.io/projected/1ca29946-9c57-462b-b53e-1bb596bfa3bb-kube-api-access-tlrh9\") pod \"multus-admission-controller-857f4d67dd-9m5ds\" (UID: \"1ca29946-9c57-462b-b53e-1bb596bfa3bb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.044430 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d42baab-089b-497d-8a9f-aea1966a66eb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sp97v\" (UID: \"4d42baab-089b-497d-8a9f-aea1966a66eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.044512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d42baab-089b-497d-8a9f-aea1966a66eb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-sp97v\" (UID: \"4d42baab-089b-497d-8a9f-aea1966a66eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.044569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-trusted-ca-bundle\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.044620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96qqt\" (UniqueName: \"kubernetes.io/projected/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-kube-api-access-96qqt\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.044697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn9xj\" (UniqueName: \"kubernetes.io/projected/ed5d8f62-5fc4-46f9-bc8b-0154549a9222-kube-api-access-kn9xj\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdt2p\" (UID: \"ed5d8f62-5fc4-46f9-bc8b-0154549a9222\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.044805 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.044894 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt2tn\" (UniqueName: \"kubernetes.io/projected/e5fc7754-5768-4bda-9534-15bd44890e2c-kube-api-access-gt2tn\") pod \"openshift-config-operator-7777fb866f-z84zq\" (UID: \"e5fc7754-5768-4bda-9534-15bd44890e2c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.044949 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhtfp\" (UniqueName: \"kubernetes.io/projected/4d42baab-089b-497d-8a9f-aea1966a66eb-kube-api-access-mhtfp\") pod \"openshift-controller-manager-operator-756b6f6bc6-sp97v\" (UID: \"4d42baab-089b-497d-8a9f-aea1966a66eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.045071 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-etcd-service-ca\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.045130 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rv5c\" (UniqueName: \"kubernetes.io/projected/e615c362-aa4f-488f-ba46-9fbb9c66e204-kube-api-access-4rv5c\") pod \"catalog-operator-68c6474976-6fnxs\" (UID: \"e615c362-aa4f-488f-ba46-9fbb9c66e204\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.045222 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-etcd-ca\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.045264 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae0463e-7748-4f59-8074-91cd1364115f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rhjr7\" (UID: \"4ae0463e-7748-4f59-8074-91cd1364115f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.046093 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.046674 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e5fc7754-5768-4bda-9534-15bd44890e2c-serving-cert\") pod \"openshift-config-operator-7777fb866f-z84zq\" (UID: \"e5fc7754-5768-4bda-9534-15bd44890e2c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.047387 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-etcd-ca\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.047466 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-trusted-ca-bundle\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.047928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4c680757-4206-4c9f-ba2b-d5443aae9f94-metrics-tls\") pod \"dns-operator-744455d44c-v9kk6\" (UID: \"4c680757-4206-4c9f-ba2b-d5443aae9f94\") " pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.047548 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-etcd-service-ca\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048015 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-secret-volume\") pod \"collect-profiles-29461830-7v4gm\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048093 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvz77\" (UniqueName: \"kubernetes.io/projected/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-kube-api-access-dvz77\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048122 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-config-volume\") pod \"collect-profiles-29461830-7v4gm\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048154 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pbv8\" (UniqueName: \"kubernetes.io/projected/0640a05a-4a45-4622-8749-88a0621fc348-kube-api-access-5pbv8\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048251 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6a26250c-db32-4acf-aa49-3a6c0d8bf017-metrics-tls\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048306 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-oauth-config\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048332 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-serving-cert\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048357 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-config\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-oauth-serving-cert\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048420 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-serving-cert\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048444 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5sb8\" (UniqueName: \"kubernetes.io/projected/40203b7c-e649-4742-b0b4-6963fb2e1aca-kube-api-access-g5sb8\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048467 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gms62\" (UniqueName: \"kubernetes.io/projected/06761c5c-371b-4593-9251-7e374633468c-kube-api-access-gms62\") pod \"migrator-59844c95c7-trj42\" (UID: \"06761c5c-371b-4593-9251-7e374633468c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048504 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m8rf\" (UniqueName: \"kubernetes.io/projected/4ae0463e-7748-4f59-8074-91cd1364115f-kube-api-access-8m8rf\") pod \"kube-storage-version-migrator-operator-b67b599dd-rhjr7\" (UID: \"4ae0463e-7748-4f59-8074-91cd1364115f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048529 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a26250c-db32-4acf-aa49-3a6c0d8bf017-trusted-ca\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d42baab-089b-497d-8a9f-aea1966a66eb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sp97v\" (UID: \"4d42baab-089b-497d-8a9f-aea1966a66eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048625 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngnnn\" (UniqueName: \"kubernetes.io/projected/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-kube-api-access-ngnnn\") pod \"collect-profiles-29461830-7v4gm\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/40203b7c-e649-4742-b0b4-6963fb2e1aca-stats-auth\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048690 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ed5d8f62-5fc4-46f9-bc8b-0154549a9222-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdt2p\" (UID: \"ed5d8f62-5fc4-46f9-bc8b-0154549a9222\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048716 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rtw8\" (UniqueName: \"kubernetes.io/projected/8c3025a2-7183-4772-b64e-97737ccbef30-kube-api-access-7rtw8\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048752 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e5fc7754-5768-4bda-9534-15bd44890e2c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z84zq\" (UID: \"e5fc7754-5768-4bda-9534-15bd44890e2c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p7jg\" (UniqueName: \"kubernetes.io/projected/4c680757-4206-4c9f-ba2b-d5443aae9f94-kube-api-access-2p7jg\") pod \"dns-operator-744455d44c-v9kk6\" (UID: \"4c680757-4206-4c9f-ba2b-d5443aae9f94\") " pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048827 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6a26250c-db32-4acf-aa49-3a6c0d8bf017-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048885 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-etcd-client\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048912 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ae0463e-7748-4f59-8074-91cd1364115f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rhjr7\" (UID: \"4ae0463e-7748-4f59-8074-91cd1364115f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.048937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40203b7c-e649-4742-b0b4-6963fb2e1aca-service-ca-bundle\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.049004 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c3025a2-7183-4772-b64e-97737ccbef30-config\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.049026 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8c3025a2-7183-4772-b64e-97737ccbef30-machine-approver-tls\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.049065 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-service-ca\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.049086 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c3025a2-7183-4772-b64e-97737ccbef30-auth-proxy-config\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.049113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e615c362-aa4f-488f-ba46-9fbb9c66e204-srv-cert\") pod \"catalog-operator-68c6474976-6fnxs\" (UID: \"e615c362-aa4f-488f-ba46-9fbb9c66e204\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.049779 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-oauth-serving-cert\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.050632 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c3025a2-7183-4772-b64e-97737ccbef30-config\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.050721 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4c680757-4206-4c9f-ba2b-d5443aae9f94-metrics-tls\") pod \"dns-operator-744455d44c-v9kk6\" (UID: \"4c680757-4206-4c9f-ba2b-d5443aae9f94\") " pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.050995 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-config\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.051459 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8c3025a2-7183-4772-b64e-97737ccbef30-auth-proxy-config\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.051866 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-serving-cert\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.052258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/e5fc7754-5768-4bda-9534-15bd44890e2c-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z84zq\" (UID: \"e5fc7754-5768-4bda-9534-15bd44890e2c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.052811 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.053078 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-service-ca\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.053646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-oauth-config\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.053690 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8c3025a2-7183-4772-b64e-97737ccbef30-machine-approver-tls\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.053973 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d42baab-089b-497d-8a9f-aea1966a66eb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-sp97v\" (UID: \"4d42baab-089b-497d-8a9f-aea1966a66eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.054275 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-etcd-client\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.056790 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-serving-cert\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.066331 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.074385 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6a26250c-db32-4acf-aa49-3a6c0d8bf017-metrics-tls\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.086506 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.121750 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.126327 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.132093 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6a26250c-db32-4acf-aa49-3a6c0d8bf017-trusted-ca\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.146247 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.166581 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.186312 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.209295 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.217862 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.226571 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.246480 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.266471 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.287498 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.296745 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/40203b7c-e649-4742-b0b4-6963fb2e1aca-stats-auth\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.306647 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.318207 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/40203b7c-e649-4742-b0b4-6963fb2e1aca-default-certificate\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.326142 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.345892 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.365939 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.375047 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ed5d8f62-5fc4-46f9-bc8b-0154549a9222-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdt2p\" (UID: \"ed5d8f62-5fc4-46f9-bc8b-0154549a9222\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.386819 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.408877 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.426796 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.445967 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.466624 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.471753 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40203b7c-e649-4742-b0b4-6963fb2e1aca-service-ca-bundle\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.488013 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.495907 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40203b7c-e649-4742-b0b4-6963fb2e1aca-metrics-certs\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.507495 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.527131 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.534623 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1ca29946-9c57-462b-b53e-1bb596bfa3bb-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9m5ds\" (UID: \"1ca29946-9c57-462b-b53e-1bb596bfa3bb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.546280 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.566986 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.586601 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.607133 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.631527 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.647773 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.665470 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.687073 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.706765 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.711414 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.712258 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.715889 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-secret-volume\") pod \"collect-profiles-29461830-7v4gm\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.718755 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e615c362-aa4f-488f-ba46-9fbb9c66e204-profile-collector-cert\") pod \"catalog-operator-68c6474976-6fnxs\" (UID: \"e615c362-aa4f-488f-ba46-9fbb9c66e204\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.728121 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.730131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-config-volume\") pod \"collect-profiles-29461830-7v4gm\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.746953 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.767287 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.787343 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.804622 4744 request.go:700] Waited for 1.019207735s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0 Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.806697 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.815086 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ae0463e-7748-4f59-8074-91cd1364115f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rhjr7\" (UID: \"4ae0463e-7748-4f59-8074-91cd1364115f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.826364 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.847682 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.857218 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ae0463e-7748-4f59-8074-91cd1364115f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rhjr7\" (UID: \"4ae0463e-7748-4f59-8074-91cd1364115f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.867093 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.887740 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.907058 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915197 4744 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915285 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.415263303 +0000 UTC m=+33.042729621 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915517 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-error: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915555 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-error podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.41554762 +0000 UTC m=+33.043013938 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-error" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-error") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915586 4744 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915604 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-cliconfig podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.415598592 +0000 UTC m=+33.043064910 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-cliconfig") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915621 4744 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915641 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config podName:5402a80f-2077-4207-81bc-cc12a2935886 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.415635323 +0000 UTC m=+33.043101641 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config") pod "route-controller-manager-6576b87f9c-h7fn9" (UID: "5402a80f-2077-4207-81bc-cc12a2935886") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915658 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-login: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915676 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.415670524 +0000 UTC m=+33.043136842 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-login" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915693 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-ocp-branding-template: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.915713 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.415707695 +0000 UTC m=+33.043174013 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-ocp-branding-template" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.916355 4744 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.916367 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.916456 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert podName:5402a80f-2077-4207-81bc-cc12a2935886 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.416427763 +0000 UTC m=+33.043894111 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert") pod "route-controller-manager-6576b87f9c-h7fn9" (UID: "5402a80f-2077-4207-81bc-cc12a2935886") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.916478 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.416468625 +0000 UTC m=+33.043934943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.917961 4744 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.918095 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-encryption-config podName:c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.418060787 +0000 UTC m=+33.045527135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-encryption-config") pod "apiserver-7bbb656c7d-5wkdx" (UID: "c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.919153 4744 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.919513 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.419491904 +0000 UTC m=+33.046958252 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.919530 4744 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.919584 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-serving-cert podName:c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.419568856 +0000 UTC m=+33.047035164 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-serving-cert") pod "apiserver-7bbb656c7d-5wkdx" (UID: "c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.921938 4744 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.921987 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-serving-ca podName:c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.42197785 +0000 UTC m=+33.049444168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-serving-ca") pod "apiserver-7bbb656c7d-5wkdx" (UID: "c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.922456 4744 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.922570 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-trusted-ca-bundle podName:c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.422549255 +0000 UTC m=+33.050015573 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-trusted-ca-bundle") pod "apiserver-7bbb656c7d-5wkdx" (UID: "c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.934395 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938467 4744 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938520 4744 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938558 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-idp-0-file-data: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938531 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-client podName:c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.438517355 +0000 UTC m=+33.065983673 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-client") pod "apiserver-7bbb656c7d-5wkdx" (UID: "c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938589 4744 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938620 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938601 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.438584627 +0000 UTC m=+33.066050945 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938706 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-idp-0-file-data podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.43867928 +0000 UTC m=+33.066145638 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-idp-0-file-data" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-idp-0-file-data") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938739 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.438721451 +0000 UTC m=+33.066187839 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: E0106 14:37:15.938823 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-client-ca podName:5402a80f-2077-4207-81bc-cc12a2935886 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:16.438754272 +0000 UTC m=+33.066220680 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-client-ca") pod "route-controller-manager-6576b87f9c-h7fn9" (UID: "5402a80f-2077-4207-81bc-cc12a2935886") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.948447 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.956612 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e615c362-aa4f-488f-ba46-9fbb9c66e204-srv-cert\") pod \"catalog-operator-68c6474976-6fnxs\" (UID: \"e615c362-aa4f-488f-ba46-9fbb9c66e204\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:15 crc kubenswrapper[4744]: I0106 14:37:15.986040 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.006469 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.026421 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.046505 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.066629 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.087138 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.127274 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.146992 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.166897 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.199872 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.208131 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.227823 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.246659 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.267674 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.286777 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.305612 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.327172 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.346219 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.366764 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.388471 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.406701 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.427643 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.447536 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.466799 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.472966 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-serving-cert\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473203 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473461 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473541 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-client\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473600 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473657 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473701 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-client-ca\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473772 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473845 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.473914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.474017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.474062 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.474098 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.474134 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.474304 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.474394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-encryption-config\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.474449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.486536 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.507398 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.526324 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.547405 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.567128 4744 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.586341 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.606979 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.629817 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.646248 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.666754 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.687130 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.706641 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.727063 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.746394 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.795849 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55gwc\" (UniqueName: \"kubernetes.io/projected/f1d1d915-9b82-4311-974c-ea3d5e58bf80-kube-api-access-55gwc\") pod \"controller-manager-879f6c89f-8pc7r\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.805373 4744 request.go:700] Waited for 1.887306409s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/serviceaccounts/machine-api-operator/token Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.817279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsj5c\" (UniqueName: \"kubernetes.io/projected/8430f458-83cd-4453-bb18-883d49bae7d9-kube-api-access-qsj5c\") pod \"authentication-operator-69f744f599-2qvhp\" (UID: \"8430f458-83cd-4453-bb18-883d49bae7d9\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.828578 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25dcd\" (UniqueName: \"kubernetes.io/projected/5e458d23-aef6-4e91-9741-46d74c403131-kube-api-access-25dcd\") pod \"machine-api-operator-5694c8668f-mpb2d\" (UID: \"5e458d23-aef6-4e91-9741-46d74c403131\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.856078 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.866013 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k6g9\" (UniqueName: \"kubernetes.io/projected/b683331c-2c30-482e-a85d-975669ae737e-kube-api-access-4k6g9\") pod \"machine-config-controller-84d6567774-zk6z4\" (UID: \"b683331c-2c30-482e-a85d-975669ae737e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.896331 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx77r\" (UniqueName: \"kubernetes.io/projected/c031a005-c650-4129-b764-55673c0e3ddd-kube-api-access-jx77r\") pod \"openshift-apiserver-operator-796bbdcf4f-7d8pp\" (UID: \"c031a005-c650-4129-b764-55673c0e3ddd\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.906296 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.917838 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt2l9\" (UniqueName: \"kubernetes.io/projected/a67f83fd-2c95-427b-afd8-bc19d5abda9e-kube-api-access-zt2l9\") pod \"apiserver-76f77b778f-qplq4\" (UID: \"a67f83fd-2c95-427b-afd8-bc19d5abda9e\") " pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.935885 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.948719 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2nh4\" (UniqueName: \"kubernetes.io/projected/87c264b6-7ed1-4b76-bc4b-a2c4e61ac543-kube-api-access-s2nh4\") pod \"cluster-samples-operator-665b6dd947-k4q45\" (UID: \"87c264b6-7ed1-4b76-bc4b-a2c4e61ac543\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.949013 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.974391 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" Jan 06 14:37:16 crc kubenswrapper[4744]: I0106 14:37:16.975216 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f8df\" (UniqueName: \"kubernetes.io/projected/26427892-f331-42fd-98a5-687d7efe8d6a-kube-api-access-9f8df\") pod \"console-operator-58897d9998-2hcpt\" (UID: \"26427892-f331-42fd-98a5-687d7efe8d6a\") " pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.009510 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq45t\" (UniqueName: \"kubernetes.io/projected/5f7af32f-7d06-4c00-b613-552287fcb5af-kube-api-access-qq45t\") pod \"downloads-7954f5f757-bbfrh\" (UID: \"5f7af32f-7d06-4c00-b613-552287fcb5af\") " pod="openshift-console/downloads-7954f5f757-bbfrh" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.033788 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.042737 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v68kx\" (UniqueName: \"kubernetes.io/projected/6a26250c-db32-4acf-aa49-3a6c0d8bf017-kube-api-access-v68kx\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.062736 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlrh9\" (UniqueName: \"kubernetes.io/projected/1ca29946-9c57-462b-b53e-1bb596bfa3bb-kube-api-access-tlrh9\") pod \"multus-admission-controller-857f4d67dd-9m5ds\" (UID: \"1ca29946-9c57-462b-b53e-1bb596bfa3bb\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.083021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96qqt\" (UniqueName: \"kubernetes.io/projected/fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526-kube-api-access-96qqt\") pod \"etcd-operator-b45778765-vnmvf\" (UID: \"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.088946 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.097690 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.111285 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn9xj\" (UniqueName: \"kubernetes.io/projected/ed5d8f62-5fc4-46f9-bc8b-0154549a9222-kube-api-access-kn9xj\") pod \"control-plane-machine-set-operator-78cbb6b69f-gdt2p\" (UID: \"ed5d8f62-5fc4-46f9-bc8b-0154549a9222\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.113872 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.124491 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.125244 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt2tn\" (UniqueName: \"kubernetes.io/projected/e5fc7754-5768-4bda-9534-15bd44890e2c-kube-api-access-gt2tn\") pod \"openshift-config-operator-7777fb866f-z84zq\" (UID: \"e5fc7754-5768-4bda-9534-15bd44890e2c\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.147672 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhtfp\" (UniqueName: \"kubernetes.io/projected/4d42baab-089b-497d-8a9f-aea1966a66eb-kube-api-access-mhtfp\") pod \"openshift-controller-manager-operator-756b6f6bc6-sp97v\" (UID: \"4d42baab-089b-497d-8a9f-aea1966a66eb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.178220 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rv5c\" (UniqueName: \"kubernetes.io/projected/e615c362-aa4f-488f-ba46-9fbb9c66e204-kube-api-access-4rv5c\") pod \"catalog-operator-68c6474976-6fnxs\" (UID: \"e615c362-aa4f-488f-ba46-9fbb9c66e204\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.180776 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvz77\" (UniqueName: \"kubernetes.io/projected/ba59ae1e-bb13-47f2-b96a-850b3270ebfb-kube-api-access-dvz77\") pod \"cluster-image-registry-operator-dc59b4c8b-h6lf7\" (UID: \"ba59ae1e-bb13-47f2-b96a-850b3270ebfb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.206787 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pbv8\" (UniqueName: \"kubernetes.io/projected/0640a05a-4a45-4622-8749-88a0621fc348-kube-api-access-5pbv8\") pod \"console-f9d7485db-6pgv4\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.232225 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.238313 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5sb8\" (UniqueName: \"kubernetes.io/projected/40203b7c-e649-4742-b0b4-6963fb2e1aca-kube-api-access-g5sb8\") pod \"router-default-5444994796-tgntc\" (UID: \"40203b7c-e649-4742-b0b4-6963fb2e1aca\") " pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.261884 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gms62\" (UniqueName: \"kubernetes.io/projected/06761c5c-371b-4593-9251-7e374633468c-kube-api-access-gms62\") pod \"migrator-59844c95c7-trj42\" (UID: \"06761c5c-371b-4593-9251-7e374633468c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.271293 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.277731 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m8rf\" (UniqueName: \"kubernetes.io/projected/4ae0463e-7748-4f59-8074-91cd1364115f-kube-api-access-8m8rf\") pod \"kube-storage-version-migrator-operator-b67b599dd-rhjr7\" (UID: \"4ae0463e-7748-4f59-8074-91cd1364115f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.287765 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.288186 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bbfrh" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.291680 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rtw8\" (UniqueName: \"kubernetes.io/projected/8c3025a2-7183-4772-b64e-97737ccbef30-kube-api-access-7rtw8\") pod \"machine-approver-56656f9798-bzgxt\" (UID: \"8c3025a2-7183-4772-b64e-97737ccbef30\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.293244 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.299920 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.307602 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.312596 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngnnn\" (UniqueName: \"kubernetes.io/projected/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-kube-api-access-ngnnn\") pod \"collect-profiles-29461830-7v4gm\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.340922 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p7jg\" (UniqueName: \"kubernetes.io/projected/4c680757-4206-4c9f-ba2b-d5443aae9f94-kube-api-access-2p7jg\") pod \"dns-operator-744455d44c-v9kk6\" (UID: \"4c680757-4206-4c9f-ba2b-d5443aae9f94\") " pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.350591 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.354673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6a26250c-db32-4acf-aa49-3a6c0d8bf017-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2zcx8\" (UID: \"6a26250c-db32-4acf-aa49-3a6c0d8bf017\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.355827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.367534 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.373194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.373262 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.391652 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.392036 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.392307 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:33.392281344 +0000 UTC m=+50.019747662 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.396237 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.399225 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.405925 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pc7r"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.409899 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.409929 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.415069 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.430585 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.434725 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mpb2d"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.434914 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.441120 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.445792 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.455395 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-client-ca\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.465278 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474122 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474191 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.474175751 +0000 UTC m=+35.101642069 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474210 4744 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474231 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-client podName:c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.474225553 +0000 UTC m=+35.101691871 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-client") pod "apiserver-7bbb656c7d-5wkdx" (UID: "c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474279 4744 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474300 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config podName:5402a80f-2077-4207-81bc-cc12a2935886 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.474295095 +0000 UTC m=+35.101761413 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config") pod "route-controller-manager-6576b87f9c-h7fn9" (UID: "5402a80f-2077-4207-81bc-cc12a2935886") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474311 4744 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474330 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-serving-cert podName:c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.474325305 +0000 UTC m=+35.101791623 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-serving-cert") pod "apiserver-7bbb656c7d-5wkdx" (UID: "c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474350 4744 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474367 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-trusted-ca-bundle podName:c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.474360626 +0000 UTC m=+35.101826944 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-trusted-ca-bundle") pod "apiserver-7bbb656c7d-5wkdx" (UID: "c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474385 4744 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.474404 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.474398107 +0000 UTC m=+35.101864425 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475655 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-login: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475691 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.475681261 +0000 UTC m=+35.103147579 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-login" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475732 4744 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-service-ca: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475755 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.475748743 +0000 UTC m=+35.103215061 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-service-ca" (UniqueName: "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475798 4744 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475819 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert podName:5402a80f-2077-4207-81bc-cc12a2935886 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.475813765 +0000 UTC m=+35.103280083 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert") pod "route-controller-manager-6576b87f9c-h7fn9" (UID: "5402a80f-2077-4207-81bc-cc12a2935886") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475846 4744 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475866 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.475860836 +0000 UTC m=+35.103327154 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync configmap cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475878 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-ocp-branding-template: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475895 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.475890927 +0000 UTC m=+35.103357245 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-ocp-branding-template" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475915 4744 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475934 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-encryption-config podName:c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.475928658 +0000 UTC m=+35.103394976 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-encryption-config") pod "apiserver-7bbb656c7d-5wkdx" (UID: "c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475947 4744 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.475991 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert podName:82c5f022-eee7-42a1-b9b3-e00c07783034 nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.475959858 +0000 UTC m=+35.103426176 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-serving-cert" (UniqueName: "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert") pod "oauth-openshift-558db77b4-9z7kv" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034") : failed to sync secret cache: timed out waiting for the condition Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.477917 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.485283 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.486288 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.493330 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.506333 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.527269 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.551469 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.566499 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 06 14:37:17 crc kubenswrapper[4744]: W0106 14:37:17.576204 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb683331c_2c30_482e_a85d_975669ae737e.slice/crio-af5f80606fdfa068316ebf722798f1f7ec68a4de7e8fecc42fbff6e2e2f1ab5c WatchSource:0}: Error finding container af5f80606fdfa068316ebf722798f1f7ec68a4de7e8fecc42fbff6e2e2f1ab5c: Status 404 returned error can't find the container with id af5f80606fdfa068316ebf722798f1f7ec68a4de7e8fecc42fbff6e2e2f1ab5c Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.587198 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.594920 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.595026 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.595093 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.595150 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.596518 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.604117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.604121 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.606102 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.609295 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.621524 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.622098 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4vhc\" (UniqueName: \"kubernetes.io/projected/5402a80f-2077-4207-81bc-cc12a2935886-kube-api-access-p4vhc\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.629571 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.647081 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.662952 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gznpj\" (UniqueName: \"kubernetes.io/projected/82c5f022-eee7-42a1-b9b3-e00c07783034-kube-api-access-gznpj\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.664328 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.667838 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.685553 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qplq4"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698745 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/271db951-8bd1-43de-a5bd-6af64a054375-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698802 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-webhook-cert\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698830 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s84v\" (UniqueName: \"kubernetes.io/projected/b3bf455e-520f-439a-a558-5f6b1feb046b-kube-api-access-6s84v\") pod \"package-server-manager-789f6589d5-m2bjv\" (UID: \"b3bf455e-520f-439a-a558-5f6b1feb046b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698848 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-proxy-tls\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698879 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698896 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-registry-certificates\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/271db951-8bd1-43de-a5bd-6af64a054375-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7jns\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-kube-api-access-d7jns\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698967 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nvqpq\" (UID: \"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.698989 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-tmpfs\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699060 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b-config\") pod \"kube-apiserver-operator-766d6c64bb-nvqpq\" (UID: \"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699088 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-images\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699121 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-trusted-ca\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699147 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3bf455e-520f-439a-a558-5f6b1feb046b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m2bjv\" (UID: \"b3bf455e-520f-439a-a558-5f6b1feb046b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699185 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bss8z\" (UniqueName: \"kubernetes.io/projected/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-kube-api-access-bss8z\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699200 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-apiservice-cert\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699226 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsz7j\" (UniqueName: \"kubernetes.io/projected/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-kube-api-access-gsz7j\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699259 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699274 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-bound-sa-token\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699294 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-registry-tls\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.699330 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nvqpq\" (UID: \"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.701919 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.20190674 +0000 UTC m=+34.829373058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.707310 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.714083 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.726536 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.732573 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.750764 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.766301 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.780134 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv8qd\" (UniqueName: \"kubernetes.io/projected/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-kube-api-access-vv8qd\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.785583 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801439 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801672 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsz7j\" (UniqueName: \"kubernetes.io/projected/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-kube-api-access-gsz7j\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801707 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801726 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-bound-sa-token\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801743 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-registry-tls\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801769 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nvqpq\" (UID: \"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801800 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/271db951-8bd1-43de-a5bd-6af64a054375-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801826 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-webhook-cert\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801841 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s84v\" (UniqueName: \"kubernetes.io/projected/b3bf455e-520f-439a-a558-5f6b1feb046b-kube-api-access-6s84v\") pod \"package-server-manager-789f6589d5-m2bjv\" (UID: \"b3bf455e-520f-439a-a558-5f6b1feb046b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801856 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-proxy-tls\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801896 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-registry-certificates\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801919 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/271db951-8bd1-43de-a5bd-6af64a054375-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7jns\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-kube-api-access-d7jns\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801958 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nvqpq\" (UID: \"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.801974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-tmpfs\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.802024 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b-config\") pod \"kube-apiserver-operator-766d6c64bb-nvqpq\" (UID: \"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.802037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-images\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.802068 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-trusted-ca\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.802083 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3bf455e-520f-439a-a558-5f6b1feb046b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m2bjv\" (UID: \"b3bf455e-520f-439a-a558-5f6b1feb046b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.802114 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bss8z\" (UniqueName: \"kubernetes.io/projected/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-kube-api-access-bss8z\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.802131 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-apiservice-cert\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.804058 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-tmpfs\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.804761 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/271db951-8bd1-43de-a5bd-6af64a054375-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.805110 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.305090467 +0000 UTC m=+34.932556825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.805752 4744 request.go:700] Waited for 1.696653296s due to client-side throttling, not priority and fairness, request: PATCH:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-879f6c89f-8pc7r/status Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.806543 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-registry-certificates\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.807484 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b-config\") pod \"kube-apiserver-operator-766d6c64bb-nvqpq\" (UID: \"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.807736 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-trusted-ca\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.807992 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-images\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.808876 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-auth-proxy-config\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.823723 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/271db951-8bd1-43de-a5bd-6af64a054375-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.827491 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-apiservice-cert\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.832947 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nvqpq\" (UID: \"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.833336 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.835817 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3bf455e-520f-439a-a558-5f6b1feb046b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m2bjv\" (UID: \"b3bf455e-520f-439a-a558-5f6b1feb046b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.837145 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-webhook-cert\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.838204 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-registry-tls\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.852125 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.858839 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-proxy-tls\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.865821 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.867317 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.887506 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.902982 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9dpx\" (UniqueName: \"kubernetes.io/projected/2661e587-fe7e-42c8-822d-812b82b15072-kube-api-access-v9dpx\") pod \"service-ca-9c57cc56f-s5h9w\" (UID: \"2661e587-fe7e-42c8-822d-812b82b15072\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903072 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-mountpoint-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c74e136e-df36-42d1-9d7a-cc3f26183998-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kbc4w\" (UID: \"c74e136e-df36-42d1-9d7a-cc3f26183998\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903108 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn5fp\" (UniqueName: \"kubernetes.io/projected/df8c0d6b-5dc0-42cc-90c0-c254f463275b-kube-api-access-cn5fp\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2661e587-fe7e-42c8-822d-812b82b15072-signing-cabundle\") pod \"service-ca-9c57cc56f-s5h9w\" (UID: \"2661e587-fe7e-42c8-822d-812b82b15072\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903176 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2661e587-fe7e-42c8-822d-812b82b15072-signing-key\") pod \"service-ca-9c57cc56f-s5h9w\" (UID: \"2661e587-fe7e-42c8-822d-812b82b15072\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903232 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kzgpw\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903248 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-registration-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903333 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903354 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2l9g\" (UniqueName: \"kubernetes.io/projected/ebecb074-406c-4d47-b8a9-356c6c2d1467-kube-api-access-f2l9g\") pod \"ingress-canary-8wjqf\" (UID: \"ebecb074-406c-4d47-b8a9-356c6c2d1467\") " pod="openshift-ingress-canary/ingress-canary-8wjqf" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903371 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-plugins-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903385 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn9nf\" (UniqueName: \"kubernetes.io/projected/bf357e72-5cc1-4710-8899-4f112347ca46-kube-api-access-qn9nf\") pod \"service-ca-operator-777779d784-fzf9s\" (UID: \"bf357e72-5cc1-4710-8899-4f112347ca46\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903412 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8375f27c-c2e0-410f-ae6a-e0339461ec55-certs\") pod \"machine-config-server-52sxd\" (UID: \"8375f27c-c2e0-410f-ae6a-e0339461ec55\") " pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903459 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d95tp\" (UniqueName: \"kubernetes.io/projected/07c8ee97-dd6b-4b05-a354-5b2250ff6709-kube-api-access-d95tp\") pod \"olm-operator-6b444d44fb-h5wzw\" (UID: \"07c8ee97-dd6b-4b05-a354-5b2250ff6709\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903493 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ebecb074-406c-4d47-b8a9-356c6c2d1467-cert\") pod \"ingress-canary-8wjqf\" (UID: \"ebecb074-406c-4d47-b8a9-356c6c2d1467\") " pod="openshift-ingress-canary/ingress-canary-8wjqf" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903567 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr4l4\" (UniqueName: \"kubernetes.io/projected/8375f27c-c2e0-410f-ae6a-e0339461ec55-kube-api-access-lr4l4\") pod \"machine-config-server-52sxd\" (UID: \"8375f27c-c2e0-410f-ae6a-e0339461ec55\") " pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903583 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c74e136e-df36-42d1-9d7a-cc3f26183998-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kbc4w\" (UID: \"c74e136e-df36-42d1-9d7a-cc3f26183998\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlspf\" (UniqueName: \"kubernetes.io/projected/2bf34d55-4b6d-4716-aca0-026cdd02aabe-kube-api-access-qlspf\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/df8c0d6b-5dc0-42cc-90c0-c254f463275b-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903737 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjrkz\" (UniqueName: \"kubernetes.io/projected/2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2-kube-api-access-gjrkz\") pod \"dns-default-6xk62\" (UID: \"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2\") " pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903747 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vnmvf"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903752 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4058a4f-d0e8-4779-9f9c-34b5e973a16f-config\") pod \"kube-controller-manager-operator-78b949d7b-f7x8h\" (UID: \"e4058a4f-d0e8-4779-9f9c-34b5e973a16f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/df8c0d6b-5dc0-42cc-90c0-c254f463275b-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.903929 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-socket-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.904188 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf357e72-5cc1-4710-8899-4f112347ca46-config\") pod \"service-ca-operator-777779d784-fzf9s\" (UID: \"bf357e72-5cc1-4710-8899-4f112347ca46\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.904299 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/07c8ee97-dd6b-4b05-a354-5b2250ff6709-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h5wzw\" (UID: \"07c8ee97-dd6b-4b05-a354-5b2250ff6709\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.904374 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf357e72-5cc1-4710-8899-4f112347ca46-serving-cert\") pod \"service-ca-operator-777779d784-fzf9s\" (UID: \"bf357e72-5cc1-4710-8899-4f112347ca46\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.904469 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4058a4f-d0e8-4779-9f9c-34b5e973a16f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-f7x8h\" (UID: \"e4058a4f-d0e8-4779-9f9c-34b5e973a16f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.904537 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-csi-data-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:17 crc kubenswrapper[4744]: E0106 14:37:17.905094 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.405082321 +0000 UTC m=+35.032548639 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.908357 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2-config-volume\") pod \"dns-default-6xk62\" (UID: \"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2\") " pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.912777 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/07c8ee97-dd6b-4b05-a354-5b2250ff6709-srv-cert\") pod \"olm-operator-6b444d44fb-h5wzw\" (UID: \"07c8ee97-dd6b-4b05-a354-5b2250ff6709\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.913078 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c74e136e-df36-42d1-9d7a-cc3f26183998-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kbc4w\" (UID: \"c74e136e-df36-42d1-9d7a-cc3f26183998\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.913134 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kzgpw\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.913267 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8375f27c-c2e0-410f-ae6a-e0339461ec55-node-bootstrap-token\") pod \"machine-config-server-52sxd\" (UID: \"8375f27c-c2e0-410f-ae6a-e0339461ec55\") " pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.916142 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2-metrics-tls\") pod \"dns-default-6xk62\" (UID: \"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2\") " pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.916209 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xldjt\" (UniqueName: \"kubernetes.io/projected/f0690b11-315a-4639-8455-d3323eb7e042-kube-api-access-xldjt\") pod \"marketplace-operator-79b997595-kzgpw\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.916232 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4058a4f-d0e8-4779-9f9c-34b5e973a16f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-f7x8h\" (UID: \"e4058a4f-d0e8-4779-9f9c-34b5e973a16f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.916281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/df8c0d6b-5dc0-42cc-90c0-c254f463275b-ready\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.918396 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.919936 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9m5ds"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.921738 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.925545 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.946317 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2qvhp"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.947691 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.948545 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.950267 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2hcpt"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.981903 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6pgv4"] Jan 06 14:37:17 crc kubenswrapper[4744]: I0106 14:37:17.987707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-bound-sa-token\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.009337 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7jns\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-kube-api-access-d7jns\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.017139 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.017257 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.517235295 +0000 UTC m=+35.144701613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.017604 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2l9g\" (UniqueName: \"kubernetes.io/projected/ebecb074-406c-4d47-b8a9-356c6c2d1467-kube-api-access-f2l9g\") pod \"ingress-canary-8wjqf\" (UID: \"ebecb074-406c-4d47-b8a9-356c6c2d1467\") " pod="openshift-ingress-canary/ingress-canary-8wjqf" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.017806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-plugins-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.018073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-plugins-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.017833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn9nf\" (UniqueName: \"kubernetes.io/projected/bf357e72-5cc1-4710-8899-4f112347ca46-kube-api-access-qn9nf\") pod \"service-ca-operator-777779d784-fzf9s\" (UID: \"bf357e72-5cc1-4710-8899-4f112347ca46\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.018130 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8375f27c-c2e0-410f-ae6a-e0339461ec55-certs\") pod \"machine-config-server-52sxd\" (UID: \"8375f27c-c2e0-410f-ae6a-e0339461ec55\") " pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.018263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d95tp\" (UniqueName: \"kubernetes.io/projected/07c8ee97-dd6b-4b05-a354-5b2250ff6709-kube-api-access-d95tp\") pod \"olm-operator-6b444d44fb-h5wzw\" (UID: \"07c8ee97-dd6b-4b05-a354-5b2250ff6709\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.018792 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ebecb074-406c-4d47-b8a9-356c6c2d1467-cert\") pod \"ingress-canary-8wjqf\" (UID: \"ebecb074-406c-4d47-b8a9-356c6c2d1467\") " pod="openshift-ingress-canary/ingress-canary-8wjqf" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr4l4\" (UniqueName: \"kubernetes.io/projected/8375f27c-c2e0-410f-ae6a-e0339461ec55-kube-api-access-lr4l4\") pod \"machine-config-server-52sxd\" (UID: \"8375f27c-c2e0-410f-ae6a-e0339461ec55\") " pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c74e136e-df36-42d1-9d7a-cc3f26183998-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kbc4w\" (UID: \"c74e136e-df36-42d1-9d7a-cc3f26183998\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019064 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlspf\" (UniqueName: \"kubernetes.io/projected/2bf34d55-4b6d-4716-aca0-026cdd02aabe-kube-api-access-qlspf\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019104 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/df8c0d6b-5dc0-42cc-90c0-c254f463275b-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019135 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjrkz\" (UniqueName: \"kubernetes.io/projected/2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2-kube-api-access-gjrkz\") pod \"dns-default-6xk62\" (UID: \"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2\") " pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4058a4f-d0e8-4779-9f9c-34b5e973a16f-config\") pod \"kube-controller-manager-operator-78b949d7b-f7x8h\" (UID: \"e4058a4f-d0e8-4779-9f9c-34b5e973a16f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019218 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/df8c0d6b-5dc0-42cc-90c0-c254f463275b-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019264 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-socket-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf357e72-5cc1-4710-8899-4f112347ca46-config\") pod \"service-ca-operator-777779d784-fzf9s\" (UID: \"bf357e72-5cc1-4710-8899-4f112347ca46\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/07c8ee97-dd6b-4b05-a354-5b2250ff6709-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h5wzw\" (UID: \"07c8ee97-dd6b-4b05-a354-5b2250ff6709\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019358 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf357e72-5cc1-4710-8899-4f112347ca46-serving-cert\") pod \"service-ca-operator-777779d784-fzf9s\" (UID: \"bf357e72-5cc1-4710-8899-4f112347ca46\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019393 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4058a4f-d0e8-4779-9f9c-34b5e973a16f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-f7x8h\" (UID: \"e4058a4f-d0e8-4779-9f9c-34b5e973a16f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019419 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-csi-data-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019444 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2-config-volume\") pod \"dns-default-6xk62\" (UID: \"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2\") " pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019467 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/07c8ee97-dd6b-4b05-a354-5b2250ff6709-srv-cert\") pod \"olm-operator-6b444d44fb-h5wzw\" (UID: \"07c8ee97-dd6b-4b05-a354-5b2250ff6709\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019490 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c74e136e-df36-42d1-9d7a-cc3f26183998-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kbc4w\" (UID: \"c74e136e-df36-42d1-9d7a-cc3f26183998\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kzgpw\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019531 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8375f27c-c2e0-410f-ae6a-e0339461ec55-node-bootstrap-token\") pod \"machine-config-server-52sxd\" (UID: \"8375f27c-c2e0-410f-ae6a-e0339461ec55\") " pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019571 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xldjt\" (UniqueName: \"kubernetes.io/projected/f0690b11-315a-4639-8455-d3323eb7e042-kube-api-access-xldjt\") pod \"marketplace-operator-79b997595-kzgpw\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4058a4f-d0e8-4779-9f9c-34b5e973a16f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-f7x8h\" (UID: \"e4058a4f-d0e8-4779-9f9c-34b5e973a16f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019614 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/df8c0d6b-5dc0-42cc-90c0-c254f463275b-ready\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019637 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2-metrics-tls\") pod \"dns-default-6xk62\" (UID: \"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2\") " pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9dpx\" (UniqueName: \"kubernetes.io/projected/2661e587-fe7e-42c8-822d-812b82b15072-kube-api-access-v9dpx\") pod \"service-ca-9c57cc56f-s5h9w\" (UID: \"2661e587-fe7e-42c8-822d-812b82b15072\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019693 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-mountpoint-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019716 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c74e136e-df36-42d1-9d7a-cc3f26183998-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kbc4w\" (UID: \"c74e136e-df36-42d1-9d7a-cc3f26183998\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019739 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn5fp\" (UniqueName: \"kubernetes.io/projected/df8c0d6b-5dc0-42cc-90c0-c254f463275b-kube-api-access-cn5fp\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019772 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2661e587-fe7e-42c8-822d-812b82b15072-signing-cabundle\") pod \"service-ca-9c57cc56f-s5h9w\" (UID: \"2661e587-fe7e-42c8-822d-812b82b15072\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019794 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2661e587-fe7e-42c8-822d-812b82b15072-signing-key\") pod \"service-ca-9c57cc56f-s5h9w\" (UID: \"2661e587-fe7e-42c8-822d-812b82b15072\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-registration-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019864 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kzgpw\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.019902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.020245 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.520230874 +0000 UTC m=+35.147697192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.020474 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-mountpoint-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.020557 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nvqpq\" (UID: \"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.023985 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/df8c0d6b-5dc0-42cc-90c0-c254f463275b-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.020879 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-csi-data-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.024476 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/df8c0d6b-5dc0-42cc-90c0-c254f463275b-ready\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.024667 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c74e136e-df36-42d1-9d7a-cc3f26183998-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kbc4w\" (UID: \"c74e136e-df36-42d1-9d7a-cc3f26183998\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.027298 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ebecb074-406c-4d47-b8a9-356c6c2d1467-cert\") pod \"ingress-canary-8wjqf\" (UID: \"ebecb074-406c-4d47-b8a9-356c6c2d1467\") " pod="openshift-ingress-canary/ingress-canary-8wjqf" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.027692 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kzgpw\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.027894 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4058a4f-d0e8-4779-9f9c-34b5e973a16f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-f7x8h\" (UID: \"e4058a4f-d0e8-4779-9f9c-34b5e973a16f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.028371 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/07c8ee97-dd6b-4b05-a354-5b2250ff6709-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h5wzw\" (UID: \"07c8ee97-dd6b-4b05-a354-5b2250ff6709\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.028460 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8375f27c-c2e0-410f-ae6a-e0339461ec55-certs\") pod \"machine-config-server-52sxd\" (UID: \"8375f27c-c2e0-410f-ae6a-e0339461ec55\") " pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.028494 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2661e587-fe7e-42c8-822d-812b82b15072-signing-cabundle\") pod \"service-ca-9c57cc56f-s5h9w\" (UID: \"2661e587-fe7e-42c8-822d-812b82b15072\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.029862 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kzgpw\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.030053 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-registration-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.030123 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/df8c0d6b-5dc0-42cc-90c0-c254f463275b-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.031398 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2bf34d55-4b6d-4716-aca0-026cdd02aabe-socket-dir\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.031484 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8375f27c-c2e0-410f-ae6a-e0339461ec55-node-bootstrap-token\") pod \"machine-config-server-52sxd\" (UID: \"8375f27c-c2e0-410f-ae6a-e0339461ec55\") " pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.031481 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2-config-volume\") pod \"dns-default-6xk62\" (UID: \"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2\") " pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.032035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf357e72-5cc1-4710-8899-4f112347ca46-config\") pod \"service-ca-operator-777779d784-fzf9s\" (UID: \"bf357e72-5cc1-4710-8899-4f112347ca46\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.032793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4058a4f-d0e8-4779-9f9c-34b5e973a16f-config\") pod \"kube-controller-manager-operator-78b949d7b-f7x8h\" (UID: \"e4058a4f-d0e8-4779-9f9c-34b5e973a16f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.034193 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf357e72-5cc1-4710-8899-4f112347ca46-serving-cert\") pod \"service-ca-operator-777779d784-fzf9s\" (UID: \"bf357e72-5cc1-4710-8899-4f112347ca46\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.034632 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2-metrics-tls\") pod \"dns-default-6xk62\" (UID: \"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2\") " pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.038746 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c74e136e-df36-42d1-9d7a-cc3f26183998-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kbc4w\" (UID: \"c74e136e-df36-42d1-9d7a-cc3f26183998\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.043927 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/07c8ee97-dd6b-4b05-a354-5b2250ff6709-srv-cert\") pod \"olm-operator-6b444d44fb-h5wzw\" (UID: \"07c8ee97-dd6b-4b05-a354-5b2250ff6709\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.044991 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2661e587-fe7e-42c8-822d-812b82b15072-signing-key\") pod \"service-ca-9c57cc56f-s5h9w\" (UID: \"2661e587-fe7e-42c8-822d-812b82b15072\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.050868 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsz7j\" (UniqueName: \"kubernetes.io/projected/2ee87c14-56de-41f8-8eff-4cdffd1fa77f-kube-api-access-gsz7j\") pod \"packageserver-d55dfcdfc-g62nk\" (UID: \"2ee87c14-56de-41f8-8eff-4cdffd1fa77f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.067066 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s84v\" (UniqueName: \"kubernetes.io/projected/b3bf455e-520f-439a-a558-5f6b1feb046b-kube-api-access-6s84v\") pod \"package-server-manager-789f6589d5-m2bjv\" (UID: \"b3bf455e-520f-439a-a558-5f6b1feb046b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.109700 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.113684 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bss8z\" (UniqueName: \"kubernetes.io/projected/d644a32a-3bfc-47d4-a8fd-90c7a99b9a39-kube-api-access-bss8z\") pod \"machine-config-operator-74547568cd-nk77z\" (UID: \"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.123998 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.140357 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.140836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2l9g\" (UniqueName: \"kubernetes.io/projected/ebecb074-406c-4d47-b8a9-356c6c2d1467-kube-api-access-f2l9g\") pod \"ingress-canary-8wjqf\" (UID: \"ebecb074-406c-4d47-b8a9-356c6c2d1467\") " pod="openshift-ingress-canary/ingress-canary-8wjqf" Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.140954 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.640932373 +0000 UTC m=+35.268398701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.147470 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn9nf\" (UniqueName: \"kubernetes.io/projected/bf357e72-5cc1-4710-8899-4f112347ca46-kube-api-access-qn9nf\") pod \"service-ca-operator-777779d784-fzf9s\" (UID: \"bf357e72-5cc1-4710-8899-4f112347ca46\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.150790 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.155759 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.157385 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bbfrh"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.157699 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.158069 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" event={"ID":"8c3025a2-7183-4772-b64e-97737ccbef30","Type":"ContainerStarted","Data":"53beb6a20f97c3b8052c695703145280550c6efd6bf2c3026894d9be31b46768"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.158111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" event={"ID":"8c3025a2-7183-4772-b64e-97737ccbef30","Type":"ContainerStarted","Data":"36163851684afc356aa85baa0438c52ca8d4d7e11e6f58e29127d664eed5e603"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.159815 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d95tp\" (UniqueName: \"kubernetes.io/projected/07c8ee97-dd6b-4b05-a354-5b2250ff6709-kube-api-access-d95tp\") pod \"olm-operator-6b444d44fb-h5wzw\" (UID: \"07c8ee97-dd6b-4b05-a354-5b2250ff6709\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.161096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" event={"ID":"b683331c-2c30-482e-a85d-975669ae737e","Type":"ContainerStarted","Data":"3118775de15a1b9115467c69f536678c68d8f51fcfd0b786dee2fd1cf79ead34"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.161128 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" event={"ID":"b683331c-2c30-482e-a85d-975669ae737e","Type":"ContainerStarted","Data":"af5f80606fdfa068316ebf722798f1f7ec68a4de7e8fecc42fbff6e2e2f1ab5c"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.163723 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" event={"ID":"a67f83fd-2c95-427b-afd8-bc19d5abda9e","Type":"ContainerStarted","Data":"1a7b0680db4b9d91b5aa1598ce0f575d741478e085e11035950a0553ac1657e1"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.165360 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" event={"ID":"5e458d23-aef6-4e91-9741-46d74c403131","Type":"ContainerStarted","Data":"6df0962e677138ac75ed4d5bd914172aba24da99387e1b31675498da92bf45d1"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.165384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" event={"ID":"5e458d23-aef6-4e91-9741-46d74c403131","Type":"ContainerStarted","Data":"4352234d35296b746f60c6cc7ea06db733b638c8567a8564e3429cf450e866cc"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.166838 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-tgntc" event={"ID":"40203b7c-e649-4742-b0b4-6963fb2e1aca","Type":"ContainerStarted","Data":"376e921f989ed531d35b999f43c911cf341964a7220153b9d4a10793c8672fd0"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.166862 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-tgntc" event={"ID":"40203b7c-e649-4742-b0b4-6963fb2e1aca","Type":"ContainerStarted","Data":"7ee1acc841390cc436a9cf594190bc6f34c4d52081b90b9475dd6d52f5621269"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.176306 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" event={"ID":"87c264b6-7ed1-4b76-bc4b-a2c4e61ac543","Type":"ContainerStarted","Data":"bb07f043c229a545bd4073197590b44a53b7b7ae2e17baa7284f1d6e33373ce6"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.177902 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" event={"ID":"e615c362-aa4f-488f-ba46-9fbb9c66e204","Type":"ContainerStarted","Data":"192359a79ced5695a951a316d8f48891c1abca36f933fc32a329f87858cd7899"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.179280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" event={"ID":"8430f458-83cd-4453-bb18-883d49bae7d9","Type":"ContainerStarted","Data":"488a52f66fa901bdb035c4ca373b68c3931a8442e98e2fc7ea6eb9eeab07e321"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.180442 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6pgv4" event={"ID":"0640a05a-4a45-4622-8749-88a0621fc348","Type":"ContainerStarted","Data":"e2a447c38ea2d3b299ee47b6b5e6c876a0de965fbd629d388eb0a9a7bdc5f6af"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.181220 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlspf\" (UniqueName: \"kubernetes.io/projected/2bf34d55-4b6d-4716-aca0-026cdd02aabe-kube-api-access-qlspf\") pod \"csi-hostpathplugin-9l4fk\" (UID: \"2bf34d55-4b6d-4716-aca0-026cdd02aabe\") " pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.184389 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" event={"ID":"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526","Type":"ContainerStarted","Data":"5a44f04793d667f29fc58dbc90e499f61f8a27c367c2b286f973ef587fb814c1"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.188707 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" event={"ID":"1ca29946-9c57-462b-b53e-1bb596bfa3bb","Type":"ContainerStarted","Data":"95290c28275b9aa7a3a62a4df03f9772f9d10aa660b72349278a11af4ff6d8d4"} Jan 06 14:37:18 crc kubenswrapper[4744]: W0106 14:37:18.191674 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ae0463e_7748_4f59_8074_91cd1364115f.slice/crio-2b14aef527a3c1aacacb8b84767e4b59ba68f3d6e7064702501a390913a9396c WatchSource:0}: Error finding container 2b14aef527a3c1aacacb8b84767e4b59ba68f3d6e7064702501a390913a9396c: Status 404 returned error can't find the container with id 2b14aef527a3c1aacacb8b84767e4b59ba68f3d6e7064702501a390913a9396c Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.194392 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2hcpt" event={"ID":"26427892-f331-42fd-98a5-687d7efe8d6a","Type":"ContainerStarted","Data":"c292f8c8bb98f413c67c05b8e489d6f746e451e759d95432ebb775d10b330652"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.194590 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.199804 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr4l4\" (UniqueName: \"kubernetes.io/projected/8375f27c-c2e0-410f-ae6a-e0339461ec55-kube-api-access-lr4l4\") pod \"machine-config-server-52sxd\" (UID: \"8375f27c-c2e0-410f-ae6a-e0339461ec55\") " pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.211034 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" event={"ID":"c031a005-c650-4129-b764-55673c0e3ddd","Type":"ContainerStarted","Data":"5ecd353ecec0b7293bed5aebd3905e59cf2bbe50bc86fffe4357dbcc893248cb"} Jan 06 14:37:18 crc kubenswrapper[4744]: W0106 14:37:18.211189 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba59ae1e_bb13_47f2_b96a_850b3270ebfb.slice/crio-cfb5496f155b10c9415e96a3c2c94f88300af52e869b35a7d3f53845330a61fc WatchSource:0}: Error finding container cfb5496f155b10c9415e96a3c2c94f88300af52e869b35a7d3f53845330a61fc: Status 404 returned error can't find the container with id cfb5496f155b10c9415e96a3c2c94f88300af52e869b35a7d3f53845330a61fc Jan 06 14:37:18 crc kubenswrapper[4744]: W0106 14:37:18.212597 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a26250c_db32_4acf_aa49_3a6c0d8bf017.slice/crio-5b9c8eaf288eae8a278f291fc76db676b32408e3d1511e2836612a54b9092901 WatchSource:0}: Error finding container 5b9c8eaf288eae8a278f291fc76db676b32408e3d1511e2836612a54b9092901: Status 404 returned error can't find the container with id 5b9c8eaf288eae8a278f291fc76db676b32408e3d1511e2836612a54b9092901 Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.218393 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" event={"ID":"f1d1d915-9b82-4311-974c-ea3d5e58bf80","Type":"ContainerStarted","Data":"b93b5e525d91494de8f362dd2451c35e648fce5a30674226d96f26ff925b6de0"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.218428 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" event={"ID":"f1d1d915-9b82-4311-974c-ea3d5e58bf80","Type":"ContainerStarted","Data":"23972bb0f2a4ae17cdc0ffcdcee4b14703555500354aff947101d609080337c0"} Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.218667 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.224998 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.225296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xldjt\" (UniqueName: \"kubernetes.io/projected/f0690b11-315a-4639-8455-d3323eb7e042-kube-api-access-xldjt\") pod \"marketplace-operator-79b997595-kzgpw\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.242311 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.242591 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.74258065 +0000 UTC m=+35.370046968 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.244085 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn5fp\" (UniqueName: \"kubernetes.io/projected/df8c0d6b-5dc0-42cc-90c0-c254f463275b-kube-api-access-cn5fp\") pod \"cni-sysctl-allowlist-ds-bd9mm\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.254145 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.264003 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.271873 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjrkz\" (UniqueName: \"kubernetes.io/projected/2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2-kube-api-access-gjrkz\") pod \"dns-default-6xk62\" (UID: \"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2\") " pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.276729 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-52sxd" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.282928 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.284991 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9dpx\" (UniqueName: \"kubernetes.io/projected/2661e587-fe7e-42c8-822d-812b82b15072-kube-api-access-v9dpx\") pod \"service-ca-9c57cc56f-s5h9w\" (UID: \"2661e587-fe7e-42c8-822d-812b82b15072\") " pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.290633 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8wjqf" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.301064 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.316258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c74e136e-df36-42d1-9d7a-cc3f26183998-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-kbc4w\" (UID: \"c74e136e-df36-42d1-9d7a-cc3f26183998\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.327068 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e4058a4f-d0e8-4779-9f9c-34b5e973a16f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-f7x8h\" (UID: \"e4058a4f-d0e8-4779-9f9c-34b5e973a16f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.343715 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.343885 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.843866748 +0000 UTC m=+35.471333066 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.344023 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.345532 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.845510171 +0000 UTC m=+35.472976489 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.387346 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.391732 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.404602 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.407303 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.411780 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.414180 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:18 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:18 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:18 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.414220 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.423926 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z84zq"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.444920 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.445370 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:18.945355071 +0000 UTC m=+35.572821389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.448437 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.448885 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.460652 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.467889 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.481206 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.491189 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.522416 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.550933 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.550966 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.550991 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551012 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551052 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551071 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551086 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-encryption-config\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551102 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551123 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-serving-cert\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551147 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551204 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551231 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-client\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.551252 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.551538 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.051499967 +0000 UTC m=+35.678966285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.553511 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.555343 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.557216 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.559798 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.560177 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.599721 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-encryption-config\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.600233 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.601791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.606950 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-etcd-client\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.607239 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.607286 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-9z7kv\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.623565 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6-serving-cert\") pod \"apiserver-7bbb656c7d-5wkdx\" (UID: \"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.627094 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-v9kk6"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.627247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert\") pod \"route-controller-manager-6576b87f9c-h7fn9\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:18 crc kubenswrapper[4744]: W0106 14:37:18.634621 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d42baab_089b_497d_8a9f_aea1966a66eb.slice/crio-e7bb94d976dbb713243b7b30739d10cadbcda86c545c365e5fa26d2fe6bd4845 WatchSource:0}: Error finding container e7bb94d976dbb713243b7b30739d10cadbcda86c545c365e5fa26d2fe6bd4845: Status 404 returned error can't find the container with id e7bb94d976dbb713243b7b30739d10cadbcda86c545c365e5fa26d2fe6bd4845 Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.654203 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.655027 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.154398377 +0000 UTC m=+35.781864685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.655069 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.655615 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.155608029 +0000 UTC m=+35.783074347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.669080 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.753105 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv"] Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.756099 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.756192 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.256176018 +0000 UTC m=+35.883642336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.767549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.767915 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.267902757 +0000 UTC m=+35.895369075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.839666 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.863445 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.870132 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.870430 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.870633 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.370619222 +0000 UTC m=+35.998085540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:18 crc kubenswrapper[4744]: I0106 14:37:18.971639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:18 crc kubenswrapper[4744]: E0106 14:37:18.972216 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.472203798 +0000 UTC m=+36.099670116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.077637 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.077958 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.577943213 +0000 UTC m=+36.205409531 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.182392 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.188458 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.688427353 +0000 UTC m=+36.315893671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.310665 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.314593 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.814569105 +0000 UTC m=+36.442035423 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.327064 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.327426 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.827412984 +0000 UTC m=+36.454879302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.370933 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9l4fk"] Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.400340 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" event={"ID":"4c680757-4206-4c9f-ba2b-d5443aae9f94","Type":"ContainerStarted","Data":"c8d223095be8deeb514a360ae3af5d71be86e2f1417c72d463b0950213c296cd"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.433718 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.434059 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:19.934043742 +0000 UTC m=+36.561510060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.441182 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:19 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:19 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:19 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.441230 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.454219 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" event={"ID":"b683331c-2c30-482e-a85d-975669ae737e","Type":"ContainerStarted","Data":"129bd278d174e91c3d4f0a122b20c90a7c470b948aa8a21b1064960e3b11d8bf"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.464658 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6pgv4" event={"ID":"0640a05a-4a45-4622-8749-88a0621fc348","Type":"ContainerStarted","Data":"757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.479739 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-52sxd" event={"ID":"8375f27c-c2e0-410f-ae6a-e0339461ec55","Type":"ContainerStarted","Data":"f4a5f4056010ed36828e46f593c9af68e5ccdc3c3e34ece2a009aa7a8bcfb75d"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.506568 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" event={"ID":"df8c0d6b-5dc0-42cc-90c0-c254f463275b","Type":"ContainerStarted","Data":"22b032a283dd099d38837b08a53cc9e5be13ebafb82f46c058dba83be9de7fef"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.528459 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" event={"ID":"bf357e72-5cc1-4710-8899-4f112347ca46","Type":"ContainerStarted","Data":"364bccfd2d9166bc3f13a64c8ba441294f6e7e20cdd5c2aa18c2baf5560df750"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.539430 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" event={"ID":"2ee87c14-56de-41f8-8eff-4cdffd1fa77f","Type":"ContainerStarted","Data":"cd978f2902188d8f212514c303a81fa322e8aa225da4c8ed097080e3c6a9ea6a"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.539938 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.540287 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.04027516 +0000 UTC m=+36.667741478 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.584485 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" event={"ID":"e615c362-aa4f-488f-ba46-9fbb9c66e204","Type":"ContainerStarted","Data":"10fa129ac99a92c4c93fc9bfa3ab0680b56a6b045d8f0d8e3fa0d993b2908c00"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.585458 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.612599 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bbfrh" event={"ID":"5f7af32f-7d06-4c00-b613-552287fcb5af","Type":"ContainerStarted","Data":"72f85b045b93bed60f0de340ca24b85b98ce38af9b722abad821e5e0b3759105"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.620903 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bbfrh" Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.636085 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42" event={"ID":"06761c5c-371b-4593-9251-7e374633468c","Type":"ContainerStarted","Data":"b6aa19c66c155eca8082780c2cfb212db567cdd9441c6c4cee95870e55aa4cf9"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.637596 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" event={"ID":"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e","Type":"ContainerStarted","Data":"a4b85543657736c7597bbcb512d89445b78e39e3537eb4acc831ec916773cc57"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.638506 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-bbfrh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.638544 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bbfrh" podUID="5f7af32f-7d06-4c00-b613-552287fcb5af" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.658242 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.658700 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.158679649 +0000 UTC m=+36.786145967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.658751 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.659955 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.159937172 +0000 UTC m=+36.787403490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.663568 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w"] Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.670078 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.688545 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ab9bf32e7a764b9d99a0d620778cc54a75698d7515ef201cb6f0f2e5f3cfc0bc"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.718225 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-6pgv4" podStartSLOduration=18.718207447 podStartE2EDuration="18.718207447s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:19.705146093 +0000 UTC m=+36.332612411" watchObservedRunningTime="2026-01-06 14:37:19.718207447 +0000 UTC m=+36.345673765" Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.766637 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.767029 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.267012872 +0000 UTC m=+36.894479190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.865827 4744 generic.go:334] "Generic (PLEG): container finished" podID="a67f83fd-2c95-427b-afd8-bc19d5abda9e" containerID="c8d36d597f089afa1dfdf823281d84f1501bcfa433724bce666170909ecd7320" exitCode=0 Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.868649 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.870225 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.37020929 +0000 UTC m=+36.997675608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.880833 4744 patch_prober.go:28] interesting pod/console-operator-58897d9998-2hcpt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.880962 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-2hcpt" podUID="26427892-f331-42fd-98a5-687d7efe8d6a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955598 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" event={"ID":"b3bf455e-520f-439a-a558-5f6b1feb046b","Type":"ContainerStarted","Data":"04cb9b825d0dbd129e2f38e5f58fdb1da38d710a9bbda28f5075aa319f7ea6a1"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955823 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" event={"ID":"87c264b6-7ed1-4b76-bc4b-a2c4e61ac543","Type":"ContainerStarted","Data":"d23adba6f2f48b7efab6d0a4225ed3a01d351e91a2e21cc1af5bbd4d04c54ee0"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955834 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" event={"ID":"8c3025a2-7183-4772-b64e-97737ccbef30","Type":"ContainerStarted","Data":"9a7471d05705ac95d1f8f19f93c071e30bd26c5222f91a1e5aa02a8279638f41"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955850 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955859 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8f1b438c436df85238b195b28482c538961231a8e4d1054fbdb8bee357f02261"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955868 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" event={"ID":"5e458d23-aef6-4e91-9741-46d74c403131","Type":"ContainerStarted","Data":"6f5f204c1aa7058b47fbe6d6bbdebd0178ee3f152319b3ca7911b5c2ec9dfaf6"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955882 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" event={"ID":"ba59ae1e-bb13-47f2-b96a-850b3270ebfb","Type":"ContainerStarted","Data":"cfb5496f155b10c9415e96a3c2c94f88300af52e869b35a7d3f53845330a61fc"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955893 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" event={"ID":"a67f83fd-2c95-427b-afd8-bc19d5abda9e","Type":"ContainerDied","Data":"c8d36d597f089afa1dfdf823281d84f1501bcfa433724bce666170909ecd7320"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2hcpt" event={"ID":"26427892-f331-42fd-98a5-687d7efe8d6a","Type":"ContainerStarted","Data":"94edf6d9ae31decac3050a1ab31d5ca3bb9bbbb9eaa6447c746a3a4863a3cf61"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955915 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq"] Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955932 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z"] Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.955943 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" event={"ID":"c031a005-c650-4129-b764-55673c0e3ddd","Type":"ContainerStarted","Data":"c799f8d1a4e0cf9a7bfaf42e016d756fcb88cbb1c78c0fae9ae4d6ffb982ce23"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.965501 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" event={"ID":"8430f458-83cd-4453-bb18-883d49bae7d9","Type":"ContainerStarted","Data":"e5c84f2d963e13dd67a1a163f6ea985e6fc20db3348d0e43da8ed54b40c8540f"} Jan 06 14:37:19 crc kubenswrapper[4744]: I0106 14:37:19.972232 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:19 crc kubenswrapper[4744]: E0106 14:37:19.973433 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.473417708 +0000 UTC m=+37.100884026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.001961 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" podStartSLOduration=18.00194341 podStartE2EDuration="18.00194341s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:19.939409263 +0000 UTC m=+36.566875601" watchObservedRunningTime="2026-01-06 14:37:20.00194341 +0000 UTC m=+36.629409728" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.002551 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" podStartSLOduration=19.002546546 podStartE2EDuration="19.002546546s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:19.994198376 +0000 UTC m=+36.621664694" watchObservedRunningTime="2026-01-06 14:37:20.002546546 +0000 UTC m=+36.630012864" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.029100 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" event={"ID":"4ae0463e-7748-4f59-8074-91cd1364115f","Type":"ContainerStarted","Data":"a634c5696eca5ac9ae032b2227ff74a03182b23a394032e080a3a80b4da7dc5a"} Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.029143 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" event={"ID":"4ae0463e-7748-4f59-8074-91cd1364115f","Type":"ContainerStarted","Data":"2b14aef527a3c1aacacb8b84767e4b59ba68f3d6e7064702501a390913a9396c"} Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.063812 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" event={"ID":"ed5d8f62-5fc4-46f9-bc8b-0154549a9222","Type":"ContainerStarted","Data":"6df77cc10afbbc771e5cb3221180b01caaf3eac0b3b845dfed93d3d106257e91"} Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.081282 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:20 crc kubenswrapper[4744]: E0106 14:37:20.086718 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.586701962 +0000 UTC m=+37.214168280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.108446 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"132ccfffbafc4a9e9456f1add307efd6e72a7bc60766330a19031115c82fc009"} Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.112829 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk6z4" podStartSLOduration=18.11281164 podStartE2EDuration="18.11281164s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.080228702 +0000 UTC m=+36.707695020" watchObservedRunningTime="2026-01-06 14:37:20.11281164 +0000 UTC m=+36.740277958" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.127962 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" event={"ID":"e5fc7754-5768-4bda-9534-15bd44890e2c","Type":"ContainerStarted","Data":"a8147fbe7aee37e917187fab47e94648c180be3c5079d914a16e9689b90e0dff"} Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.139394 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" event={"ID":"4d42baab-089b-497d-8a9f-aea1966a66eb","Type":"ContainerStarted","Data":"e7bb94d976dbb713243b7b30739d10cadbcda86c545c365e5fa26d2fe6bd4845"} Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.141494 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" event={"ID":"fd4e3ef7-953e-4ba6-b9a3-dfbdcf0c5526","Type":"ContainerStarted","Data":"b591b665fd386da4948ad8690fe78cac97b0db4a3b90b4518517cfdee6648b6b"} Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.145283 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.151204 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-tgntc" podStartSLOduration=19.151186341 podStartE2EDuration="19.151186341s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.1504192 +0000 UTC m=+36.777885518" watchObservedRunningTime="2026-01-06 14:37:20.151186341 +0000 UTC m=+36.778652659" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.181436 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" event={"ID":"6a26250c-db32-4acf-aa49-3a6c0d8bf017","Type":"ContainerStarted","Data":"5b9c8eaf288eae8a278f291fc76db676b32408e3d1511e2836612a54b9092901"} Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.182110 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:20 crc kubenswrapper[4744]: E0106 14:37:20.183111 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.683095501 +0000 UTC m=+37.310561819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.285948 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.296728 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vnmvf" podStartSLOduration=19.296707084 podStartE2EDuration="19.296707084s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.236687873 +0000 UTC m=+36.864154191" watchObservedRunningTime="2026-01-06 14:37:20.296707084 +0000 UTC m=+36.924173392" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.297489 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-2hcpt" podStartSLOduration=19.297483194 podStartE2EDuration="19.297483194s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.2863247 +0000 UTC m=+36.913791018" watchObservedRunningTime="2026-01-06 14:37:20.297483194 +0000 UTC m=+36.924949512" Jan 06 14:37:20 crc kubenswrapper[4744]: E0106 14:37:20.299045 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.799033915 +0000 UTC m=+37.426500233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.388601 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.388734 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:20 crc kubenswrapper[4744]: E0106 14:37:20.389040 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.889022304 +0000 UTC m=+37.516488632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.429678 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:20 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:20 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:20 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.431851 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.433989 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-bbfrh" podStartSLOduration=19.433969808 podStartE2EDuration="19.433969808s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.408357573 +0000 UTC m=+37.035823891" watchObservedRunningTime="2026-01-06 14:37:20.433969808 +0000 UTC m=+37.061436126" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.443206 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4e45835e-b6fd-4d25-ad1a-c5d9b9221cde-metrics-certs\") pod \"network-metrics-daemon-f9m2v\" (UID: \"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde\") " pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.461408 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-6xk62"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.489596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:20 crc kubenswrapper[4744]: E0106 14:37:20.489930 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:20.989920012 +0000 UTC m=+37.617386330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.514064 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8wjqf"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.539082 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-f9m2v" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.590538 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.591967 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rhjr7" podStartSLOduration=18.591947029 podStartE2EDuration="18.591947029s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.542860306 +0000 UTC m=+37.170326644" watchObservedRunningTime="2026-01-06 14:37:20.591947029 +0000 UTC m=+37.219413347" Jan 06 14:37:20 crc kubenswrapper[4744]: E0106 14:37:20.592657 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.092631577 +0000 UTC m=+37.720097895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.597329 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.609588 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgpw"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.611986 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-mpb2d" podStartSLOduration=18.611968676 podStartE2EDuration="18.611968676s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.611291008 +0000 UTC m=+37.238757326" watchObservedRunningTime="2026-01-06 14:37:20.611968676 +0000 UTC m=+37.239434994" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.625670 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-s5h9w"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.640934 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c8458"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.676836 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.684140 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" podStartSLOduration=19.684117276 podStartE2EDuration="19.684117276s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.675429848 +0000 UTC m=+37.302896166" watchObservedRunningTime="2026-01-06 14:37:20.684117276 +0000 UTC m=+37.311583594" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.690327 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.691918 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c8458"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.729668 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.732731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-utilities\") pod \"certified-operators-c8458\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.732796 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jthl\" (UniqueName: \"kubernetes.io/projected/ead06eb7-e94b-400a-b568-1871fea3e807-kube-api-access-2jthl\") pod \"certified-operators-c8458\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.732840 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.732932 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-catalog-content\") pod \"certified-operators-c8458\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: E0106 14:37:20.733338 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.233325903 +0000 UTC m=+37.860792211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.790121 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.798578 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2qvhp" podStartSLOduration=19.798562411 podStartE2EDuration="19.798562411s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.778257626 +0000 UTC m=+37.405723944" watchObservedRunningTime="2026-01-06 14:37:20.798562411 +0000 UTC m=+37.426028729" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.824135 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bzgxt" podStartSLOduration=19.824116354 podStartE2EDuration="19.824116354s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.815575199 +0000 UTC m=+37.443041517" watchObservedRunningTime="2026-01-06 14:37:20.824116354 +0000 UTC m=+37.451582672" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.826227 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4cwvb"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.827123 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.829911 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.836074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.836241 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-catalog-content\") pod \"certified-operators-c8458\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.836303 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-utilities\") pod \"certified-operators-c8458\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.836326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jthl\" (UniqueName: \"kubernetes.io/projected/ead06eb7-e94b-400a-b568-1871fea3e807-kube-api-access-2jthl\") pod \"certified-operators-c8458\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: E0106 14:37:20.836835 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.336821068 +0000 UTC m=+37.964287386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.837151 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-catalog-content\") pod \"certified-operators-c8458\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.837380 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-utilities\") pod \"certified-operators-c8458\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.844410 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9z7kv"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.853678 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4cwvb"] Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.863041 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7d8pp" podStartSLOduration=19.863027479 podStartE2EDuration="19.863027479s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:20.862484044 +0000 UTC m=+37.489950362" watchObservedRunningTime="2026-01-06 14:37:20.863027479 +0000 UTC m=+37.490493797" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.878053 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jthl\" (UniqueName: \"kubernetes.io/projected/ead06eb7-e94b-400a-b568-1871fea3e807-kube-api-access-2jthl\") pod \"certified-operators-c8458\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.938876 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-catalog-content\") pod \"community-operators-4cwvb\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.939210 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-utilities\") pod \"community-operators-4cwvb\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.939255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:20 crc kubenswrapper[4744]: I0106 14:37:20.939300 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98h5j\" (UniqueName: \"kubernetes.io/projected/14d7cfd7-181f-4d79-8951-27d300d49a8f-kube-api-access-98h5j\") pod \"community-operators-4cwvb\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:20 crc kubenswrapper[4744]: E0106 14:37:20.939579 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.439567805 +0000 UTC m=+38.067034123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.002676 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j999h"] Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.003585 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.034535 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j999h"] Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.037987 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.042085 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.042477 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98h5j\" (UniqueName: \"kubernetes.io/projected/14d7cfd7-181f-4d79-8951-27d300d49a8f-kube-api-access-98h5j\") pod \"community-operators-4cwvb\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.042506 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tfgc\" (UniqueName: \"kubernetes.io/projected/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-kube-api-access-2tfgc\") pod \"certified-operators-j999h\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.042542 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-catalog-content\") pod \"community-operators-4cwvb\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.044456 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-utilities\") pod \"certified-operators-j999h\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.044509 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.544494448 +0000 UTC m=+38.171960766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.044780 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-catalog-content\") pod \"community-operators-4cwvb\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.045063 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-utilities\") pod \"community-operators-4cwvb\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.045095 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-catalog-content\") pod \"certified-operators-j999h\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.045142 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.045461 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-utilities\") pod \"community-operators-4cwvb\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.045721 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.54571294 +0000 UTC m=+38.173179258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.101813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98h5j\" (UniqueName: \"kubernetes.io/projected/14d7cfd7-181f-4d79-8951-27d300d49a8f-kube-api-access-98h5j\") pod \"community-operators-4cwvb\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.147075 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.147232 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.647213414 +0000 UTC m=+38.274679732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.147450 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-utilities\") pod \"certified-operators-j999h\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.147498 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-catalog-content\") pod \"certified-operators-j999h\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.147523 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.147551 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tfgc\" (UniqueName: \"kubernetes.io/projected/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-kube-api-access-2tfgc\") pod \"certified-operators-j999h\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.148145 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-utilities\") pod \"certified-operators-j999h\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.148418 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-catalog-content\") pod \"certified-operators-j999h\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.148622 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.648614911 +0000 UTC m=+38.276081229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.177951 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.203542 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tfgc\" (UniqueName: \"kubernetes.io/projected/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-kube-api-access-2tfgc\") pod \"certified-operators-j999h\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.206639 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m4xw2"] Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.215643 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.220041 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4xw2"] Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.245847 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" event={"ID":"1ca29946-9c57-462b-b53e-1bb596bfa3bb","Type":"ContainerStarted","Data":"247c4c4fe5f5fe8a1b6aa60dc65148a8902a5f78768e08ebcfe236d7c73b5b11"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.245883 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" event={"ID":"1ca29946-9c57-462b-b53e-1bb596bfa3bb","Type":"ContainerStarted","Data":"729a7978d5252131184b0e3bff544f2500b85a32bfcb344835701aa7fb6cb4e7"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.248712 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.249384 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.749359844 +0000 UTC m=+38.376826162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.256696 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-catalog-content\") pod \"community-operators-m4xw2\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.256754 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.256800 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49v7g\" (UniqueName: \"kubernetes.io/projected/1d41f99a-8352-472d-9011-e9c0259abf7b-kube-api-access-49v7g\") pod \"community-operators-m4xw2\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.256889 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-utilities\") pod \"community-operators-m4xw2\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.257328 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.757312064 +0000 UTC m=+38.384778372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.296080 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-52sxd" event={"ID":"8375f27c-c2e0-410f-ae6a-e0339461ec55","Type":"ContainerStarted","Data":"cac75441a6607952fb7b1ec7e8b755a71d763e5e9da51683ae7cfb5f8ff14619"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.297351 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-9m5ds" podStartSLOduration=19.297334408 podStartE2EDuration="19.297334408s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.294489123 +0000 UTC m=+37.921955451" watchObservedRunningTime="2026-01-06 14:37:21.297334408 +0000 UTC m=+37.924800726" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.308009 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" event={"ID":"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b","Type":"ContainerStarted","Data":"50b99c181d2c57fb79ee5bb90d12ab66b089d22613516a3a309eedfd1d5ad0f7"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.326193 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" event={"ID":"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39","Type":"ContainerStarted","Data":"236fe302a6faca39bd6587614247c61653ff50761565ab04579d0f01d513d6c6"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.326247 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" event={"ID":"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39","Type":"ContainerStarted","Data":"f4d32c220e37b1df61c3146b776d445e3ece262dacb75584e31f9cee3c8a56d5"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.374466 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.374873 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49v7g\" (UniqueName: \"kubernetes.io/projected/1d41f99a-8352-472d-9011-e9c0259abf7b-kube-api-access-49v7g\") pod \"community-operators-m4xw2\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.374960 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-utilities\") pod \"community-operators-m4xw2\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.375015 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-catalog-content\") pod \"community-operators-m4xw2\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.375720 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-catalog-content\") pod \"community-operators-m4xw2\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.375792 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.875777264 +0000 UTC m=+38.503243582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.376859 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-utilities\") pod \"community-operators-m4xw2\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.391343 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.417833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"06fe0defca7568f05af8e2752e1b1272b564e0202eb419ecd4c8861137a63b27"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.427073 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:21 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:21 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:21 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.427145 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.434990 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49v7g\" (UniqueName: \"kubernetes.io/projected/1d41f99a-8352-472d-9011-e9c0259abf7b-kube-api-access-49v7g\") pod \"community-operators-m4xw2\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.457489 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-52sxd" podStartSLOduration=7.457468966 podStartE2EDuration="7.457468966s" podCreationTimestamp="2026-01-06 14:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.33274217 +0000 UTC m=+37.960208488" watchObservedRunningTime="2026-01-06 14:37:21.457468966 +0000 UTC m=+38.084935284" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.487846 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" event={"ID":"bf357e72-5cc1-4710-8899-4f112347ca46","Type":"ContainerStarted","Data":"b3f6de2b041cff4a0c2bfad39c0281ebd0dfa764b2bb7b5c46c4c8d52a51a107"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.489196 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.490351 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:21.990340101 +0000 UTC m=+38.617806419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.513260 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" event={"ID":"c74e136e-df36-42d1-9d7a-cc3f26183998","Type":"ContainerStarted","Data":"c3ab5d798da5bba17883fcd460b98c410b664881dcd466aee19e588efdc4234b"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.513300 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" event={"ID":"c74e136e-df36-42d1-9d7a-cc3f26183998","Type":"ContainerStarted","Data":"272c82c3c8ba5521ad3774cf137ac03199864d9cc276b34f1bcc580d91e83b2a"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.522694 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzf9s" podStartSLOduration=19.522677553 podStartE2EDuration="19.522677553s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.521891812 +0000 UTC m=+38.149358130" watchObservedRunningTime="2026-01-06 14:37:21.522677553 +0000 UTC m=+38.150143871" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.538269 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" event={"ID":"2661e587-fe7e-42c8-822d-812b82b15072","Type":"ContainerStarted","Data":"ddb2624fc3cfb48df644e467a12c6a8bb2f2bb8e14b0b05762ac7f4e3c2edbe7"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.561459 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" event={"ID":"ed5d8f62-5fc4-46f9-bc8b-0154549a9222","Type":"ContainerStarted","Data":"538f2297514e1f6cd6a7274ebc8c8938c1e84cbfed54c9045fa56b121dc949ad"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.562814 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.587835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" event={"ID":"5402a80f-2077-4207-81bc-cc12a2935886","Type":"ContainerStarted","Data":"a12a4514dd4bf050664ad0fd076c0f28d36ea039177c4930d7560123bfcfd24d"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.587921 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gdt2p" podStartSLOduration=19.587910961 podStartE2EDuration="19.587910961s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.587563672 +0000 UTC m=+38.215029990" watchObservedRunningTime="2026-01-06 14:37:21.587910961 +0000 UTC m=+38.215377279" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.588401 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-kbc4w" podStartSLOduration=19.588394114 podStartE2EDuration="19.588394114s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.558350843 +0000 UTC m=+38.185817151" watchObservedRunningTime="2026-01-06 14:37:21.588394114 +0000 UTC m=+38.215860432" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.590957 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.596787 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:22.096751734 +0000 UTC m=+38.724218052 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.627445 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" event={"ID":"07c8ee97-dd6b-4b05-a354-5b2250ff6709","Type":"ContainerStarted","Data":"9faeb91197842b8390ea72b61956d388b755de8baa264087b4f9e32b621410da"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.648199 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" event={"ID":"4c680757-4206-4c9f-ba2b-d5443aae9f94","Type":"ContainerStarted","Data":"fa8a8bc343baaa5f23edfd9853e35df7875766fadc23f27cac8e1bf66ddf7de8"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.659003 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" event={"ID":"2bf34d55-4b6d-4716-aca0-026cdd02aabe","Type":"ContainerStarted","Data":"67e617dc3672066507fdb9cfc9cf82adad61c5ec58458b5500f3557f2df85fb5"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.693024 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.694181 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:22.1941696 +0000 UTC m=+38.821635918 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.728325 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3079847c2987504f8b6aeb6326caf6b488904a78d25eb30d45f02959a5620bf7"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.728364 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.731420 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-f9m2v"] Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.745930 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" event={"ID":"e4058a4f-d0e8-4779-9f9c-34b5e973a16f","Type":"ContainerStarted","Data":"57ef0927d1cc8b4249662f14613d644c1d75ae37c2ae1517d468e0affc876130"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.750275 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" event={"ID":"2ee87c14-56de-41f8-8eff-4cdffd1fa77f","Type":"ContainerStarted","Data":"9c3a04af746f55686a928eb40fa20e6aa8f7e5f728c3ef05493b7918aef41424"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.751607 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.768604 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" podStartSLOduration=19.76859125 podStartE2EDuration="19.76859125s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.767417219 +0000 UTC m=+38.394883537" watchObservedRunningTime="2026-01-06 14:37:21.76859125 +0000 UTC m=+38.396057568" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.774762 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" event={"ID":"6a26250c-db32-4acf-aa49-3a6c0d8bf017","Type":"ContainerStarted","Data":"18a79f6b284e8f404eb9e070d69ca963543b150c4d35cc349d5e62440bd05bb8"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.775057 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" event={"ID":"6a26250c-db32-4acf-aa49-3a6c0d8bf017","Type":"ContainerStarted","Data":"8ced5c2ce407094797854b5cf161b889a630897ec999a82ae575342c1e8540e0"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.778312 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h6lf7" event={"ID":"ba59ae1e-bb13-47f2-b96a-850b3270ebfb","Type":"ContainerStarted","Data":"92301eee07ef664804f1f81344b41f131ba2808f5efad2834135df8c63818a03"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.791255 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" event={"ID":"4d42baab-089b-497d-8a9f-aea1966a66eb","Type":"ContainerStarted","Data":"14ea7730fb3b69f6499d14303abb84944154ec6d3de1f2e4c1a16f58bbdeac91"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.793634 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.794763 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:22.294748679 +0000 UTC m=+38.922214997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.796035 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6xk62" event={"ID":"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2","Type":"ContainerStarted","Data":"760904ba8e427f042872eac69f6bee52a26f3cdd58ee9da14414f8357ee9f3a1"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.797129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8wjqf" event={"ID":"ebecb074-406c-4d47-b8a9-356c6c2d1467","Type":"ContainerStarted","Data":"de8b7ce2f52e4633605b5620eedaa04cd62192bfe16999e062d3821d2c8ed865"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.804650 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42" event={"ID":"06761c5c-371b-4593-9251-7e374633468c","Type":"ContainerStarted","Data":"332ff822a5574cae4a62f18e3c7bcb2062ef5e2d7229917951ff8c169f82bc08"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.804707 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42" event={"ID":"06761c5c-371b-4593-9251-7e374633468c","Type":"ContainerStarted","Data":"320cc178071c2313ef66570595119ad3733a4a994b970cdee87222ee6680af7d"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.863370 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bbfrh" event={"ID":"5f7af32f-7d06-4c00-b613-552287fcb5af","Type":"ContainerStarted","Data":"d4c701479d285b13e01ae7b4e73da63857fe57239f428d0bb76116256735d49d"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.864660 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-bbfrh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.864697 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bbfrh" podUID="5f7af32f-7d06-4c00-b613-552287fcb5af" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.885905 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2zcx8" podStartSLOduration=20.88589102 podStartE2EDuration="20.88589102s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.840274498 +0000 UTC m=+38.467740816" watchObservedRunningTime="2026-01-06 14:37:21.88589102 +0000 UTC m=+38.513357328" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.897427 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" event={"ID":"df8c0d6b-5dc0-42cc-90c0-c254f463275b","Type":"ContainerStarted","Data":"cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.898141 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.898721 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:21 crc kubenswrapper[4744]: E0106 14:37:21.903976 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:22.403961746 +0000 UTC m=+39.031428064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.911296 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-trj42" podStartSLOduration=19.911275568 podStartE2EDuration="19.911275568s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.887104102 +0000 UTC m=+38.514570420" watchObservedRunningTime="2026-01-06 14:37:21.911275568 +0000 UTC m=+38.538741896" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.911681 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-sp97v" podStartSLOduration=20.911676429 podStartE2EDuration="20.911676429s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.911262938 +0000 UTC m=+38.538729256" watchObservedRunningTime="2026-01-06 14:37:21.911676429 +0000 UTC m=+38.539142747" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.951712 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" event={"ID":"f0690b11-315a-4639-8455-d3323eb7e042","Type":"ContainerStarted","Data":"7774f1f0e040ab54031de788f52146dd8c7625f1210aeac6284127fce585cdc5"} Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.952787 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.955775 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" podStartSLOduration=7.95576244 podStartE2EDuration="7.95576244s" podCreationTimestamp="2026-01-06 14:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:21.935570538 +0000 UTC m=+38.563036856" watchObservedRunningTime="2026-01-06 14:37:21.95576244 +0000 UTC m=+38.583228758" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.961584 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kzgpw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.961628 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" podUID="f0690b11-315a-4639-8455-d3323eb7e042" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Jan 06 14:37:21 crc kubenswrapper[4744]: I0106 14:37:21.973771 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e52250f0486623080a4d858237b63b09b6b3dcbe65d1fb039fb6d138fc533a2a"} Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:21.999957 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.001022 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:22.501005852 +0000 UTC m=+39.128472180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.067687 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" podStartSLOduration=20.067657227 podStartE2EDuration="20.067657227s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:22.026733719 +0000 UTC m=+38.654200037" watchObservedRunningTime="2026-01-06 14:37:22.067657227 +0000 UTC m=+38.695123545" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.089476 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" event={"ID":"87c264b6-7ed1-4b76-bc4b-a2c4e61ac543","Type":"ContainerStarted","Data":"eadcac8b52b878dbf4b5a40444278132030f0bd4ce75ddcc53686e68f7bf2923"} Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.093103 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c8458"] Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.105806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.106898 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:22.60688207 +0000 UTC m=+39.234348378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.129485 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k4q45" podStartSLOduration=21.129469815 podStartE2EDuration="21.129469815s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:22.127956735 +0000 UTC m=+38.755423053" watchObservedRunningTime="2026-01-06 14:37:22.129469815 +0000 UTC m=+38.756936123" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.173710 4744 generic.go:334] "Generic (PLEG): container finished" podID="e5fc7754-5768-4bda-9534-15bd44890e2c" containerID="ee9ac9445bc19343aa9cd9b8f5820f4f533955dc23510025d923d4ad8024ffd3" exitCode=0 Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.174552 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" event={"ID":"e5fc7754-5768-4bda-9534-15bd44890e2c","Type":"ContainerStarted","Data":"7f65535d733f3d7ca3566dc57b0d5e22c8b220eb1587c2fab9e798050d9123ad"} Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.174574 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" event={"ID":"e5fc7754-5768-4bda-9534-15bd44890e2c","Type":"ContainerDied","Data":"ee9ac9445bc19343aa9cd9b8f5820f4f533955dc23510025d923d4ad8024ffd3"} Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.174605 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.206889 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.226714 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:22.726687226 +0000 UTC m=+39.354153544 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.233068 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" podStartSLOduration=21.233052874 podStartE2EDuration="21.233052874s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:22.231092392 +0000 UTC m=+38.858558700" watchObservedRunningTime="2026-01-06 14:37:22.233052874 +0000 UTC m=+38.860519192" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.308618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.309684 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:22.809672442 +0000 UTC m=+39.437138760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.312990 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" event={"ID":"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6","Type":"ContainerStarted","Data":"fc312e4d858f8d152c64c3fc647dd5bfed4ceedcf1e6d71cfc2b5a843d1eeeb6"} Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.356406 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" event={"ID":"82c5f022-eee7-42a1-b9b3-e00c07783034","Type":"ContainerStarted","Data":"2755a17385d2b7d13e8fcb5b0e112663bf80b345942916b6bb75d37131b42f94"} Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.418828 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.419494 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:22.919469214 +0000 UTC m=+39.546935532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.422655 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" event={"ID":"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e","Type":"ContainerStarted","Data":"16c5ca1f08ffad2f6c4a5f3a7a25e8fcb6bd34d9398d5bed71783bfdd8854ea8"} Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.423721 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:22 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:22 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:22 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.423758 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.434281 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.492179 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" podStartSLOduration=20.492149718 podStartE2EDuration="20.492149718s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:22.490951356 +0000 UTC m=+39.118417674" watchObservedRunningTime="2026-01-06 14:37:22.492149718 +0000 UTC m=+39.119616036" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.508237 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" event={"ID":"b3bf455e-520f-439a-a558-5f6b1feb046b","Type":"ContainerStarted","Data":"098ecafdfdf2a468c2f0b0d24878c68b280ac9dbcb13bd1465e9e4ad7f9884aa"} Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.512734 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.516601 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m4xw2"] Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.521050 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.521391 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.021379578 +0000 UTC m=+39.648845886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.558246 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-2hcpt" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.587500 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" podStartSLOduration=20.587478109 podStartE2EDuration="20.587478109s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:22.580273349 +0000 UTC m=+39.207739677" watchObservedRunningTime="2026-01-06 14:37:22.587478109 +0000 UTC m=+39.214944427" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.629067 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.630917 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.130883702 +0000 UTC m=+39.758350020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.638678 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.639683 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j999h"] Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.644476 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.14446091 +0000 UTC m=+39.771927228 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.709811 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-g62nk" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.716904 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4cwvb"] Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.743131 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.743493 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.243476798 +0000 UTC m=+39.870943116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.802950 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4hxcm"] Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.815142 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.819970 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.822052 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hxcm"] Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.844255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.844597 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.344586111 +0000 UTC m=+39.972052429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.948912 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.949150 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jh2lp\" (UniqueName: \"kubernetes.io/projected/6fe6e90c-e807-441c-b130-15a7368b87f0-kube-api-access-jh2lp\") pod \"redhat-marketplace-4hxcm\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.949207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-utilities\") pod \"redhat-marketplace-4hxcm\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:22 crc kubenswrapper[4744]: I0106 14:37:22.949240 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-catalog-content\") pod \"redhat-marketplace-4hxcm\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:22 crc kubenswrapper[4744]: E0106 14:37:22.949346 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.44933295 +0000 UTC m=+40.076799268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.050471 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-utilities\") pod \"redhat-marketplace-4hxcm\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.050525 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-catalog-content\") pod \"redhat-marketplace-4hxcm\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.050568 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.050596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jh2lp\" (UniqueName: \"kubernetes.io/projected/6fe6e90c-e807-441c-b130-15a7368b87f0-kube-api-access-jh2lp\") pod \"redhat-marketplace-4hxcm\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.051504 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.55148817 +0000 UTC m=+40.178954488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.051574 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-catalog-content\") pod \"redhat-marketplace-4hxcm\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.051663 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-utilities\") pod \"redhat-marketplace-4hxcm\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.092818 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jh2lp\" (UniqueName: \"kubernetes.io/projected/6fe6e90c-e807-441c-b130-15a7368b87f0-kube-api-access-jh2lp\") pod \"redhat-marketplace-4hxcm\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.151253 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.151588 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.651572456 +0000 UTC m=+40.279038774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.163495 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.205351 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fv9qm"] Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.213923 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.232637 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv9qm"] Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.253079 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.253438 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.753425349 +0000 UTC m=+40.380891667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.353694 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.353818 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.853790553 +0000 UTC m=+40.481256861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.353946 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.354007 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcvmd\" (UniqueName: \"kubernetes.io/projected/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-kube-api-access-wcvmd\") pod \"redhat-marketplace-fv9qm\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.354031 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-catalog-content\") pod \"redhat-marketplace-fv9qm\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.354069 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-utilities\") pod \"redhat-marketplace-fv9qm\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.354306 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.854283365 +0000 UTC m=+40.481749683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.414676 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:23 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:23 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:23 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.414879 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.454996 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.455269 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.955253695 +0000 UTC m=+40.582720013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.455380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.455416 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcvmd\" (UniqueName: \"kubernetes.io/projected/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-kube-api-access-wcvmd\") pod \"redhat-marketplace-fv9qm\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.455436 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-catalog-content\") pod \"redhat-marketplace-fv9qm\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.455464 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-utilities\") pod \"redhat-marketplace-fv9qm\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.455783 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-utilities\") pod \"redhat-marketplace-fv9qm\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.455991 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:23.955983474 +0000 UTC m=+40.583449792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.456361 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-catalog-content\") pod \"redhat-marketplace-fv9qm\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.481006 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcvmd\" (UniqueName: \"kubernetes.io/projected/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-kube-api-access-wcvmd\") pod \"redhat-marketplace-fv9qm\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.521032 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" event={"ID":"82c5f022-eee7-42a1-b9b3-e00c07783034","Type":"ContainerStarted","Data":"93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.521624 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.525619 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8wjqf" event={"ID":"ebecb074-406c-4d47-b8a9-356c6c2d1467","Type":"ContainerStarted","Data":"96922cea1628becdac414f1d996059ffeba45284041da5b360c5f1aa94fbb61b"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.527727 4744 generic.go:334] "Generic (PLEG): container finished" podID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerID="491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e" exitCode=0 Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.527850 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j999h" event={"ID":"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf","Type":"ContainerDied","Data":"491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.527959 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j999h" event={"ID":"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf","Type":"ContainerStarted","Data":"f3d39d2f6b25f93b055c1c8de6492a2e44703a12949b3f82b331050d3fcae806"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.529222 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.531466 4744 generic.go:334] "Generic (PLEG): container finished" podID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerID="c4209be2450e3939a345ed77b5b447158f39954aa943c4d1e0bd23ca4a8d1c3a" exitCode=0 Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.531646 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cwvb" event={"ID":"14d7cfd7-181f-4d79-8951-27d300d49a8f","Type":"ContainerDied","Data":"c4209be2450e3939a345ed77b5b447158f39954aa943c4d1e0bd23ca4a8d1c3a"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.531670 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cwvb" event={"ID":"14d7cfd7-181f-4d79-8951-27d300d49a8f","Type":"ContainerStarted","Data":"0fa264b620a099372f580860981c19bc14169722a645de9882d9f5f7322972a9"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.538206 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-f9m2v" event={"ID":"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde","Type":"ContainerStarted","Data":"35521f591fe81a824e952f259dcc1e7275c9e42955f3703a2ba68c8c4fe43b2d"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.538234 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-f9m2v" event={"ID":"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde","Type":"ContainerStarted","Data":"0ad2d1b55dc0f660b9fa1bd96b8dd2c653050b37ab37001343e565cd97d77c82"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.540329 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" event={"ID":"5402a80f-2077-4207-81bc-cc12a2935886","Type":"ContainerStarted","Data":"229a0808bcdf0ebfd4f50e30b9f4ab9973979e575ad9fa32f22baea299c3b9e4"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.540775 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.570968 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.571298 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:24.071284641 +0000 UTC m=+40.698750959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.580371 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" event={"ID":"f0690b11-315a-4639-8455-d3323eb7e042","Type":"ContainerStarted","Data":"a7c178b0865fc44d01d8d02f6a72c64a84c203e52857b42d4d324229547e5614"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.599742 4744 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kzgpw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.599793 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" podUID="f0690b11-315a-4639-8455-d3323eb7e042" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.600208 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.607674 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" podStartSLOduration=22.607656208999998 podStartE2EDuration="22.607656209s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:23.58075586 +0000 UTC m=+40.208222178" watchObservedRunningTime="2026-01-06 14:37:23.607656209 +0000 UTC m=+40.235122527" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.639908 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-bd9mm"] Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.650680 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" event={"ID":"a67f83fd-2c95-427b-afd8-bc19d5abda9e","Type":"ContainerStarted","Data":"44e6ff98e2e9d56089b155660dac8e63b63083777a885fff41444168e72eb9cd"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.650739 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" event={"ID":"a67f83fd-2c95-427b-afd8-bc19d5abda9e","Type":"ContainerStarted","Data":"c58275067698417c43cfc20c5116dfe4d1884a02662f3d4aaef5e2ab2e2ea49b"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.657598 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" podStartSLOduration=21.657583584 podStartE2EDuration="21.657583584s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:23.656465945 +0000 UTC m=+40.283932253" watchObservedRunningTime="2026-01-06 14:37:23.657583584 +0000 UTC m=+40.285049902" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.676375 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.678091 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:24.178075814 +0000 UTC m=+40.805542132 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.685219 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" event={"ID":"b3bf455e-520f-439a-a558-5f6b1feb046b","Type":"ContainerStarted","Data":"c5f5830b1fa0886f541203c765031ebbaf6d6cd5297985fd15bd52fa1b4f2f8b"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.740507 4744 generic.go:334] "Generic (PLEG): container finished" podID="ead06eb7-e94b-400a-b568-1871fea3e807" containerID="9d6778973ec5c041f7fa6f29df88279be671a5ecd0f5bee8f4ed8f9697996e93" exitCode=0 Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.745057 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.745100 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8458" event={"ID":"ead06eb7-e94b-400a-b568-1871fea3e807","Type":"ContainerDied","Data":"9d6778973ec5c041f7fa6f29df88279be671a5ecd0f5bee8f4ed8f9697996e93"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.745119 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8458" event={"ID":"ead06eb7-e94b-400a-b568-1871fea3e807","Type":"ContainerStarted","Data":"eba2ea840fc97fd2f00f616943413ccf0f41f571ce13d91dd17a2cd6d3794639"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.763516 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" event={"ID":"e4058a4f-d0e8-4779-9f9c-34b5e973a16f","Type":"ContainerStarted","Data":"96ff5f385f9ba647fccbf81df7caf153e714882e0920eeda9bdec880ebf817be"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.770401 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-8wjqf" podStartSLOduration=9.770387155 podStartE2EDuration="9.770387155s" podCreationTimestamp="2026-01-06 14:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:23.709242925 +0000 UTC m=+40.336709243" watchObservedRunningTime="2026-01-06 14:37:23.770387155 +0000 UTC m=+40.397853473" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.777957 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.778800 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:24.278784206 +0000 UTC m=+40.906250524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.813313 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" event={"ID":"a5bcdd3c-ec3b-401c-a18c-f7c02aeaec6b","Type":"ContainerStarted","Data":"62dbfbfe221e0ab2ff5d8714cc699234058ba8d176bceaac3ed991f13a2b6772"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.817348 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4bh4q"] Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.818314 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:23 crc kubenswrapper[4744]: W0106 14:37:23.820062 4744 reflector.go:561] object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh": failed to list *v1.Secret: secrets "redhat-operators-dockercfg-ct8rh" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.820272 4744 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-ct8rh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"redhat-operators-dockercfg-ct8rh\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.836589 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6xk62" event={"ID":"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2","Type":"ContainerStarted","Data":"d80c44b532191778af274e0c9337ff5f8673b80ac0d3e1d8214ab78cc998f89b"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.840323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" event={"ID":"2661e587-fe7e-42c8-822d-812b82b15072","Type":"ContainerStarted","Data":"20f2f1baab20271381f4f38c290c52dae34e0886c9a3380bf152d3954ebbf739"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.852472 4744 generic.go:334] "Generic (PLEG): container finished" podID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerID="395f98bd0eac1daa06856550ce62970d259464d9e5bd0c8308807d150879896e" exitCode=0 Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.852533 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4xw2" event={"ID":"1d41f99a-8352-472d-9011-e9c0259abf7b","Type":"ContainerDied","Data":"395f98bd0eac1daa06856550ce62970d259464d9e5bd0c8308807d150879896e"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.852561 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4xw2" event={"ID":"1d41f99a-8352-472d-9011-e9c0259abf7b","Type":"ContainerStarted","Data":"59f4c5028f5391f4cc262c987ca6c785281778a81d20abdca3572e3dd00d0cf6"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.855340 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bh4q"] Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.879929 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:23 crc kubenswrapper[4744]: E0106 14:37:23.881791 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:24.381778689 +0000 UTC m=+41.009245007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.893056 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nvqpq" podStartSLOduration=21.893040556 podStartE2EDuration="21.893040556s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:23.891890085 +0000 UTC m=+40.519356403" watchObservedRunningTime="2026-01-06 14:37:23.893040556 +0000 UTC m=+40.520506874" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.923721 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" event={"ID":"07c8ee97-dd6b-4b05-a354-5b2250ff6709","Type":"ContainerStarted","Data":"7f7709d5070925362727c7e9a176fba771a144400f026d6435e3ebe4ebe9b5b3"} Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.924396 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.955417 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" Jan 06 14:37:23 crc kubenswrapper[4744]: I0106 14:37:23.984687 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" event={"ID":"4c680757-4206-4c9f-ba2b-d5443aae9f94","Type":"ContainerStarted","Data":"cc80a32ff45b1f1f75f80136a24a6d549aa4d510dd9163965e4ac45fe41635e5"} Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.004573 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-69twj"] Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.006197 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.015776 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" event={"ID":"2bf34d55-4b6d-4716-aca0-026cdd02aabe","Type":"ContainerStarted","Data":"adad2c9df553c024f3ef9064ea0185ba85fc4fbca68bcc2e04239185c967afcd"} Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.025840 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.026042 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-utilities\") pod \"redhat-operators-69twj\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.055644 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhbt7\" (UniqueName: \"kubernetes.io/projected/6cbba207-bda4-4f95-a8f1-d003787cb231-kube-api-access-jhbt7\") pod \"redhat-operators-69twj\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.055707 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-catalog-content\") pod \"redhat-operators-4bh4q\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.055751 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-catalog-content\") pod \"redhat-operators-69twj\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.055792 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68bnt\" (UniqueName: \"kubernetes.io/projected/408386a1-c1c3-4b95-8bce-4c60433159a0-kube-api-access-68bnt\") pod \"redhat-operators-4bh4q\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.055837 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-utilities\") pod \"redhat-operators-4bh4q\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: E0106 14:37:24.058478 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:24.558453491 +0000 UTC m=+41.185919809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.061302 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hxcm"] Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.074741 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69twj"] Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.078328 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" event={"ID":"d644a32a-3bfc-47d4-a8fd-90c7a99b9a39","Type":"ContainerStarted","Data":"b881e2eeb483f94326ac68fc84a869a008eed1669c4190e8d147e807da0cf79b"} Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.078941 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" podStartSLOduration=23.078921121 podStartE2EDuration="23.078921121s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:24.016196698 +0000 UTC m=+40.643663016" watchObservedRunningTime="2026-01-06 14:37:24.078921121 +0000 UTC m=+40.706387429" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.079821 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-f7x8h" podStartSLOduration=22.079815524 podStartE2EDuration="22.079815524s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:24.061662146 +0000 UTC m=+40.689128464" watchObservedRunningTime="2026-01-06 14:37:24.079815524 +0000 UTC m=+40.707281842" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.100215 4744 generic.go:334] "Generic (PLEG): container finished" podID="c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6" containerID="92ee90b795f0c30810acc85b34298c39a4f55f987bda69e15d31c173f5825c54" exitCode=0 Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.101392 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" event={"ID":"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6","Type":"ContainerStarted","Data":"156519354e48249f105945fe68c93b1bcf07792f2739c9a6756716e9752c2cd6"} Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.101422 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" event={"ID":"c9678d67-4d5c-43d6-97b2-2a6f1dcc27c6","Type":"ContainerDied","Data":"92ee90b795f0c30810acc85b34298c39a4f55f987bda69e15d31c173f5825c54"} Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.109206 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-bbfrh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.109275 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bbfrh" podUID="5f7af32f-7d06-4c00-b613-552287fcb5af" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.110024 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z84zq" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.159043 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-utilities\") pod \"redhat-operators-69twj\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.159422 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhbt7\" (UniqueName: \"kubernetes.io/projected/6cbba207-bda4-4f95-a8f1-d003787cb231-kube-api-access-jhbt7\") pod \"redhat-operators-69twj\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.159507 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-catalog-content\") pod \"redhat-operators-4bh4q\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.159582 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-catalog-content\") pod \"redhat-operators-69twj\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.159674 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68bnt\" (UniqueName: \"kubernetes.io/projected/408386a1-c1c3-4b95-8bce-4c60433159a0-kube-api-access-68bnt\") pod \"redhat-operators-4bh4q\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.159740 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-utilities\") pod \"redhat-operators-4bh4q\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.159812 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.165207 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-utilities\") pod \"redhat-operators-69twj\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.165394 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-catalog-content\") pod \"redhat-operators-4bh4q\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: E0106 14:37:24.175537 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:24.675522075 +0000 UTC m=+41.302988393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.177024 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-utilities\") pod \"redhat-operators-4bh4q\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.182280 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhbt7\" (UniqueName: \"kubernetes.io/projected/6cbba207-bda4-4f95-a8f1-d003787cb231-kube-api-access-jhbt7\") pod \"redhat-operators-69twj\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.196054 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-v9kk6" podStartSLOduration=23.196037085 podStartE2EDuration="23.196037085s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:24.192330188 +0000 UTC m=+40.819796506" watchObservedRunningTime="2026-01-06 14:37:24.196037085 +0000 UTC m=+40.823503403" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.209209 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-catalog-content\") pod \"redhat-operators-69twj\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.227269 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68bnt\" (UniqueName: \"kubernetes.io/projected/408386a1-c1c3-4b95-8bce-4c60433159a0-kube-api-access-68bnt\") pod \"redhat-operators-4bh4q\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.235857 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-nk77z" podStartSLOduration=22.235844814 podStartE2EDuration="22.235844814s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:24.233512932 +0000 UTC m=+40.860979250" watchObservedRunningTime="2026-01-06 14:37:24.235844814 +0000 UTC m=+40.863311132" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.264692 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:24 crc kubenswrapper[4744]: E0106 14:37:24.265083 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:24.765065853 +0000 UTC m=+41.392532171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.280362 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-s5h9w" podStartSLOduration=22.280348596 podStartE2EDuration="22.280348596s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:24.279081703 +0000 UTC m=+40.906548021" watchObservedRunningTime="2026-01-06 14:37:24.280348596 +0000 UTC m=+40.907814914" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.335394 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h5wzw" podStartSLOduration=22.335378495 podStartE2EDuration="22.335378495s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:24.333732702 +0000 UTC m=+40.961199020" watchObservedRunningTime="2026-01-06 14:37:24.335378495 +0000 UTC m=+40.962844813" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.367708 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:24 crc kubenswrapper[4744]: E0106 14:37:24.367978 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:24.867968254 +0000 UTC m=+41.495434572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.371080 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" podStartSLOduration=22.371071125 podStartE2EDuration="22.371071125s" podCreationTimestamp="2026-01-06 14:37:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:24.368707903 +0000 UTC m=+40.996174221" watchObservedRunningTime="2026-01-06 14:37:24.371071125 +0000 UTC m=+40.998537443" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.405584 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv9qm"] Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.415268 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:24 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:24 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:24 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.415306 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.470194 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:24 crc kubenswrapper[4744]: E0106 14:37:24.470784 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:24.970760631 +0000 UTC m=+41.598226969 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.492134 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.512678 4744 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.573898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:24 crc kubenswrapper[4744]: E0106 14:37:24.574191 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:25.074180295 +0000 UTC m=+41.701646613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.674959 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:24 crc kubenswrapper[4744]: E0106 14:37:24.676106 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-06 14:37:25.176090039 +0000 UTC m=+41.803556357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.689800 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.697266 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.697675 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.780980 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:24 crc kubenswrapper[4744]: E0106 14:37:24.781337 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-06 14:37:25.281324681 +0000 UTC m=+41.908790999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8ltv4" (UID: "271db951-8bd1-43de-a5bd-6af64a054375") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.790559 4744 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-06T14:37:24.512713526Z","Handler":null,"Name":""} Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.798702 4744 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.798736 4744 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.881618 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.960112 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.984803 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:24 crc kubenswrapper[4744]: I0106 14:37:24.998555 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-69twj"] Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.005263 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.005329 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.031122 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bh4q"] Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.146734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-f9m2v" event={"ID":"4e45835e-b6fd-4d25-ad1a-c5d9b9221cde","Type":"ContainerStarted","Data":"a60dbe278c1387c378712b32f9ed37e53cd8367a42f9a3d5fd3bf6bc6abd4653"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.150070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" event={"ID":"2bf34d55-4b6d-4716-aca0-026cdd02aabe","Type":"ContainerStarted","Data":"957e266a9c377845d2511fe758e8e55497ed72a9a84799d8c5452cdb151d4975"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.150091 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" event={"ID":"2bf34d55-4b6d-4716-aca0-026cdd02aabe","Type":"ContainerStarted","Data":"6761034926b26bb22f0834d1c0b1a69bf53746dd1b85967071852bca70c4ede1"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.156292 4744 generic.go:334] "Generic (PLEG): container finished" podID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerID="5625d7cac6383b53c66ad4cd0c6a3470229ca31d4696074fa954b0c65522dde6" exitCode=0 Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.156566 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv9qm" event={"ID":"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe","Type":"ContainerDied","Data":"5625d7cac6383b53c66ad4cd0c6a3470229ca31d4696074fa954b0c65522dde6"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.156934 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv9qm" event={"ID":"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe","Type":"ContainerStarted","Data":"dbe15818a0e63f9e6998d0dca78bd5410a481811f0f4db4c5c2697858393c7b9"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.173062 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-f9m2v" podStartSLOduration=24.173043498 podStartE2EDuration="24.173043498s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:25.167119552 +0000 UTC m=+41.794585870" watchObservedRunningTime="2026-01-06 14:37:25.173043498 +0000 UTC m=+41.800509816" Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.173525 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-6xk62" event={"ID":"2fb5d2a5-0ed7-4d87-aa33-9c8738013bf2","Type":"ContainerStarted","Data":"11c3a50b90caa0819cfd769ccbe05711171e69b4524df7198ba0b06bfa18df14"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.173611 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.179077 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bh4q" event={"ID":"408386a1-c1c3-4b95-8bce-4c60433159a0","Type":"ContainerStarted","Data":"72a8d03499b533e3150127bb2e0e21cbf159a8a4d79e5760e326a06ed473a67c"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.181531 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69twj" event={"ID":"6cbba207-bda4-4f95-a8f1-d003787cb231","Type":"ContainerStarted","Data":"f967f280b5a6dad613e78688c20fa20a567e138e768b7cb4241437bebe32aab2"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.184872 4744 generic.go:334] "Generic (PLEG): container finished" podID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerID="56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb" exitCode=0 Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.186538 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hxcm" event={"ID":"6fe6e90c-e807-441c-b130-15a7368b87f0","Type":"ContainerDied","Data":"56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.186565 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hxcm" event={"ID":"6fe6e90c-e807-441c-b130-15a7368b87f0","Type":"ContainerStarted","Data":"d5f78c308bd69589e5535d1f5f1b82f57e3784e9238c57f04e821fd57f94e787"} Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.188084 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" gracePeriod=30 Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.193358 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.246736 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8ltv4\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.269870 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-6xk62" podStartSLOduration=11.269854168 podStartE2EDuration="11.269854168s" podCreationTimestamp="2026-01-06 14:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:25.264889678 +0000 UTC m=+41.892355996" watchObservedRunningTime="2026-01-06 14:37:25.269854168 +0000 UTC m=+41.897320486" Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.413849 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:25 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:25 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:25 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.414197 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.476284 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.700130 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8ltv4"] Jan 06 14:37:25 crc kubenswrapper[4744]: I0106 14:37:25.716506 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.202246 4744 generic.go:334] "Generic (PLEG): container finished" podID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerID="afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30" exitCode=0 Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.202553 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bh4q" event={"ID":"408386a1-c1c3-4b95-8bce-4c60433159a0","Type":"ContainerDied","Data":"afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30"} Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.204376 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" event={"ID":"271db951-8bd1-43de-a5bd-6af64a054375","Type":"ContainerStarted","Data":"d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada"} Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.204607 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" event={"ID":"271db951-8bd1-43de-a5bd-6af64a054375","Type":"ContainerStarted","Data":"29b6d1c7ba553bc633153a925b3ced27825bb33e39038f14f38108c0c3046325"} Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.204934 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.208997 4744 generic.go:334] "Generic (PLEG): container finished" podID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerID="15b52fb24e6d33d4e904e1c5359b148cf27973de6b1ea034f7e0004c2b646ca0" exitCode=0 Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.209072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69twj" event={"ID":"6cbba207-bda4-4f95-a8f1-d003787cb231","Type":"ContainerDied","Data":"15b52fb24e6d33d4e904e1c5359b148cf27973de6b1ea034f7e0004c2b646ca0"} Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.212329 4744 generic.go:334] "Generic (PLEG): container finished" podID="3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e" containerID="16c5ca1f08ffad2f6c4a5f3a7a25e8fcb6bd34d9398d5bed71783bfdd8854ea8" exitCode=0 Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.212420 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" event={"ID":"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e","Type":"ContainerDied","Data":"16c5ca1f08ffad2f6c4a5f3a7a25e8fcb6bd34d9398d5bed71783bfdd8854ea8"} Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.222028 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" event={"ID":"2bf34d55-4b6d-4716-aca0-026cdd02aabe","Type":"ContainerStarted","Data":"3e99c6ab55ebdceb682d6b9a6eed7ba70fbe1da697170621d74c1c9a4a66bc44"} Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.288302 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" podStartSLOduration=25.288283022 podStartE2EDuration="25.288283022s" podCreationTimestamp="2026-01-06 14:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:26.27983034 +0000 UTC m=+42.907296668" watchObservedRunningTime="2026-01-06 14:37:26.288283022 +0000 UTC m=+42.915749330" Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.327172 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" podStartSLOduration=12.327140366 podStartE2EDuration="12.327140366s" podCreationTimestamp="2026-01-06 14:37:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:26.324370403 +0000 UTC m=+42.951836721" watchObservedRunningTime="2026-01-06 14:37:26.327140366 +0000 UTC m=+42.954606684" Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.413529 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:26 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:26 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:26 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.413578 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.949993 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.950052 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:26 crc kubenswrapper[4744]: I0106 14:37:26.961505 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.242449 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-qplq4" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.273658 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.273694 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.289146 4744 patch_prober.go:28] interesting pod/console-f9d7485db-6pgv4 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.289229 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6pgv4" podUID="0640a05a-4a45-4622-8749-88a0621fc348" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.290583 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-bbfrh container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.290664 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bbfrh" podUID="5f7af32f-7d06-4c00-b613-552287fcb5af" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.291169 4744 patch_prober.go:28] interesting pod/downloads-7954f5f757-bbfrh container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" start-of-body= Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.291227 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bbfrh" podUID="5f7af32f-7d06-4c00-b613-552287fcb5af" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.13:8080/\": dial tcp 10.217.0.13:8080: connect: connection refused" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.410632 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.412691 4744 patch_prober.go:28] interesting pod/router-default-5444994796-tgntc container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 06 14:37:27 crc kubenswrapper[4744]: [-]has-synced failed: reason withheld Jan 06 14:37:27 crc kubenswrapper[4744]: [+]process-running ok Jan 06 14:37:27 crc kubenswrapper[4744]: healthz check failed Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.412749 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-tgntc" podUID="40203b7c-e649-4742-b0b4-6963fb2e1aca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.719682 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.846581 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-config-volume\") pod \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.846646 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngnnn\" (UniqueName: \"kubernetes.io/projected/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-kube-api-access-ngnnn\") pod \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.846679 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-secret-volume\") pod \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\" (UID: \"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e\") " Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.847971 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-config-volume" (OuterVolumeSpecName: "config-volume") pod "3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e" (UID: "3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.854416 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e" (UID: "3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.872684 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-kube-api-access-ngnnn" (OuterVolumeSpecName: "kube-api-access-ngnnn") pod "3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e" (UID: "3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e"). InnerVolumeSpecName "kube-api-access-ngnnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.938033 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 06 14:37:27 crc kubenswrapper[4744]: E0106 14:37:27.941151 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e" containerName="collect-profiles" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.941199 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e" containerName="collect-profiles" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.941328 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e" containerName="collect-profiles" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.941716 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.944053 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.948340 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.948360 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.948369 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngnnn\" (UniqueName: \"kubernetes.io/projected/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e-kube-api-access-ngnnn\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.958514 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 06 14:37:27 crc kubenswrapper[4744]: I0106 14:37:27.964455 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.050424 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0274d836-1692-4b10-b5f7-6282136d3ed6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0274d836-1692-4b10-b5f7-6282136d3ed6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.050537 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0274d836-1692-4b10-b5f7-6282136d3ed6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0274d836-1692-4b10-b5f7-6282136d3ed6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.151980 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0274d836-1692-4b10-b5f7-6282136d3ed6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0274d836-1692-4b10-b5f7-6282136d3ed6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.152074 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0274d836-1692-4b10-b5f7-6282136d3ed6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0274d836-1692-4b10-b5f7-6282136d3ed6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.152168 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0274d836-1692-4b10-b5f7-6282136d3ed6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0274d836-1692-4b10-b5f7-6282136d3ed6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.177993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0274d836-1692-4b10-b5f7-6282136d3ed6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0274d836-1692-4b10-b5f7-6282136d3ed6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.270523 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.286555 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.289524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm" event={"ID":"3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e","Type":"ContainerDied","Data":"a4b85543657736c7597bbcb512d89445b78e39e3537eb4acc831ec916773cc57"} Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.289563 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4b85543657736c7597bbcb512d89445b78e39e3537eb4acc831ec916773cc57" Jan 06 14:37:28 crc kubenswrapper[4744]: E0106 14:37:28.312452 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:28 crc kubenswrapper[4744]: E0106 14:37:28.337543 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:28 crc kubenswrapper[4744]: E0106 14:37:28.357657 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:28 crc kubenswrapper[4744]: E0106 14:37:28.357707 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerName="kube-multus-additional-cni-plugins" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.415583 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.421566 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-tgntc" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.782488 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.871223 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.871258 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:28 crc kubenswrapper[4744]: I0106 14:37:28.892117 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:29 crc kubenswrapper[4744]: I0106 14:37:29.297647 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0274d836-1692-4b10-b5f7-6282136d3ed6","Type":"ContainerStarted","Data":"6770111fc238577be7fce47f9658ff20515f1024f459db6277e9d0c162f0c4a1"} Jan 06 14:37:29 crc kubenswrapper[4744]: I0106 14:37:29.303245 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-5wkdx" Jan 06 14:37:29 crc kubenswrapper[4744]: I0106 14:37:29.960050 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:29 crc kubenswrapper[4744]: I0106 14:37:29.960454 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.018835 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.309871 4744 generic.go:334] "Generic (PLEG): container finished" podID="0274d836-1692-4b10-b5f7-6282136d3ed6" containerID="0740ae9c71bf345675d4e103041edaf0feecc7ba214fe16d6a34898b3ccdc903" exitCode=0 Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.310568 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0274d836-1692-4b10-b5f7-6282136d3ed6","Type":"ContainerDied","Data":"0740ae9c71bf345675d4e103041edaf0feecc7ba214fe16d6a34898b3ccdc903"} Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.786177 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.786881 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.790281 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.790444 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.795396 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.930925 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:30 crc kubenswrapper[4744]: I0106 14:37:30.931418 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:31 crc kubenswrapper[4744]: I0106 14:37:31.037947 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:31 crc kubenswrapper[4744]: I0106 14:37:31.038004 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:31 crc kubenswrapper[4744]: I0106 14:37:31.038393 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:31 crc kubenswrapper[4744]: I0106 14:37:31.059552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:31 crc kubenswrapper[4744]: I0106 14:37:31.154735 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:33 crc kubenswrapper[4744]: I0106 14:37:33.526362 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-6xk62" Jan 06 14:37:37 crc kubenswrapper[4744]: I0106 14:37:37.298817 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-bbfrh" Jan 06 14:37:37 crc kubenswrapper[4744]: I0106 14:37:37.331836 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:37 crc kubenswrapper[4744]: I0106 14:37:37.337258 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:37:37 crc kubenswrapper[4744]: I0106 14:37:37.820145 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 06 14:37:37 crc kubenswrapper[4744]: I0106 14:37:37.835578 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 06 14:37:38 crc kubenswrapper[4744]: E0106 14:37:38.269402 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:38 crc kubenswrapper[4744]: E0106 14:37:38.271657 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:38 crc kubenswrapper[4744]: E0106 14:37:38.274598 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:38 crc kubenswrapper[4744]: E0106 14:37:38.274655 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerName="kube-multus-additional-cni-plugins" Jan 06 14:37:43 crc kubenswrapper[4744]: I0106 14:37:43.226692 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pc7r"] Jan 06 14:37:43 crc kubenswrapper[4744]: I0106 14:37:43.227267 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" podUID="f1d1d915-9b82-4311-974c-ea3d5e58bf80" containerName="controller-manager" containerID="cri-o://b93b5e525d91494de8f362dd2451c35e648fce5a30674226d96f26ff925b6de0" gracePeriod=30 Jan 06 14:37:43 crc kubenswrapper[4744]: I0106 14:37:43.306062 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=6.306042935 podStartE2EDuration="6.306042935s" podCreationTimestamp="2026-01-06 14:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:43.257424394 +0000 UTC m=+59.884890712" watchObservedRunningTime="2026-01-06 14:37:43.306042935 +0000 UTC m=+59.933509253" Jan 06 14:37:43 crc kubenswrapper[4744]: I0106 14:37:43.306815 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9"] Jan 06 14:37:43 crc kubenswrapper[4744]: I0106 14:37:43.307034 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" podUID="5402a80f-2077-4207-81bc-cc12a2935886" containerName="route-controller-manager" containerID="cri-o://229a0808bcdf0ebfd4f50e30b9f4ab9973979e575ad9fa32f22baea299c3b9e4" gracePeriod=30 Jan 06 14:37:45 crc kubenswrapper[4744]: I0106 14:37:45.481422 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:37:45 crc kubenswrapper[4744]: I0106 14:37:45.739341 4744 generic.go:334] "Generic (PLEG): container finished" podID="5402a80f-2077-4207-81bc-cc12a2935886" containerID="229a0808bcdf0ebfd4f50e30b9f4ab9973979e575ad9fa32f22baea299c3b9e4" exitCode=0 Jan 06 14:37:45 crc kubenswrapper[4744]: I0106 14:37:45.739428 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" event={"ID":"5402a80f-2077-4207-81bc-cc12a2935886","Type":"ContainerDied","Data":"229a0808bcdf0ebfd4f50e30b9f4ab9973979e575ad9fa32f22baea299c3b9e4"} Jan 06 14:37:45 crc kubenswrapper[4744]: I0106 14:37:45.741232 4744 generic.go:334] "Generic (PLEG): container finished" podID="f1d1d915-9b82-4311-974c-ea3d5e58bf80" containerID="b93b5e525d91494de8f362dd2451c35e648fce5a30674226d96f26ff925b6de0" exitCode=0 Jan 06 14:37:45 crc kubenswrapper[4744]: I0106 14:37:45.741289 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" event={"ID":"f1d1d915-9b82-4311-974c-ea3d5e58bf80","Type":"ContainerDied","Data":"b93b5e525d91494de8f362dd2451c35e648fce5a30674226d96f26ff925b6de0"} Jan 06 14:37:46 crc kubenswrapper[4744]: I0106 14:37:46.857569 4744 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8pc7r container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Jan 06 14:37:46 crc kubenswrapper[4744]: I0106 14:37:46.857640 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" podUID="f1d1d915-9b82-4311-974c-ea3d5e58bf80" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Jan 06 14:37:48 crc kubenswrapper[4744]: E0106 14:37:48.267030 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:48 crc kubenswrapper[4744]: E0106 14:37:48.268270 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:48 crc kubenswrapper[4744]: E0106 14:37:48.269620 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:48 crc kubenswrapper[4744]: E0106 14:37:48.269653 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerName="kube-multus-additional-cni-plugins" Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.451720 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.645835 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0274d836-1692-4b10-b5f7-6282136d3ed6-kubelet-dir\") pod \"0274d836-1692-4b10-b5f7-6282136d3ed6\" (UID: \"0274d836-1692-4b10-b5f7-6282136d3ed6\") " Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.645947 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0274d836-1692-4b10-b5f7-6282136d3ed6-kube-api-access\") pod \"0274d836-1692-4b10-b5f7-6282136d3ed6\" (UID: \"0274d836-1692-4b10-b5f7-6282136d3ed6\") " Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.645982 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0274d836-1692-4b10-b5f7-6282136d3ed6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0274d836-1692-4b10-b5f7-6282136d3ed6" (UID: "0274d836-1692-4b10-b5f7-6282136d3ed6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.646202 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0274d836-1692-4b10-b5f7-6282136d3ed6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.654131 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0274d836-1692-4b10-b5f7-6282136d3ed6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0274d836-1692-4b10-b5f7-6282136d3ed6" (UID: "0274d836-1692-4b10-b5f7-6282136d3ed6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.747299 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0274d836-1692-4b10-b5f7-6282136d3ed6-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.760750 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0274d836-1692-4b10-b5f7-6282136d3ed6","Type":"ContainerDied","Data":"6770111fc238577be7fce47f9658ff20515f1024f459db6277e9d0c162f0c4a1"} Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.760790 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6770111fc238577be7fce47f9658ff20515f1024f459db6277e9d0c162f0c4a1" Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.760823 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.841077 4744 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-h7fn9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.841495 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" podUID="5402a80f-2077-4207-81bc-cc12a2935886" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 06 14:37:48 crc kubenswrapper[4744]: I0106 14:37:48.869012 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 06 14:37:48 crc kubenswrapper[4744]: W0106 14:37:48.881207 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4dc8108a_af68_4f28_88eb_00bbb0ba134e.slice/crio-1a3dc7ccb02d57b14df9e7753550c7fe68f3c79842848d5f5db397f74d4a3d99 WatchSource:0}: Error finding container 1a3dc7ccb02d57b14df9e7753550c7fe68f3c79842848d5f5db397f74d4a3d99: Status 404 returned error can't find the container with id 1a3dc7ccb02d57b14df9e7753550c7fe68f3c79842848d5f5db397f74d4a3d99 Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.524919 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.562987 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-client-ca\") pod \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.563063 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-config\") pod \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.563119 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-proxy-ca-bundles\") pod \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.563184 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1d1d915-9b82-4311-974c-ea3d5e58bf80-serving-cert\") pod \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.563217 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55gwc\" (UniqueName: \"kubernetes.io/projected/f1d1d915-9b82-4311-974c-ea3d5e58bf80-kube-api-access-55gwc\") pod \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\" (UID: \"f1d1d915-9b82-4311-974c-ea3d5e58bf80\") " Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.565452 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-config" (OuterVolumeSpecName: "config") pod "f1d1d915-9b82-4311-974c-ea3d5e58bf80" (UID: "f1d1d915-9b82-4311-974c-ea3d5e58bf80"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.565499 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-client-ca" (OuterVolumeSpecName: "client-ca") pod "f1d1d915-9b82-4311-974c-ea3d5e58bf80" (UID: "f1d1d915-9b82-4311-974c-ea3d5e58bf80"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.565674 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f1d1d915-9b82-4311-974c-ea3d5e58bf80" (UID: "f1d1d915-9b82-4311-974c-ea3d5e58bf80"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.587241 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1d1d915-9b82-4311-974c-ea3d5e58bf80-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f1d1d915-9b82-4311-974c-ea3d5e58bf80" (UID: "f1d1d915-9b82-4311-974c-ea3d5e58bf80"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.592050 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1d1d915-9b82-4311-974c-ea3d5e58bf80-kube-api-access-55gwc" (OuterVolumeSpecName: "kube-api-access-55gwc") pod "f1d1d915-9b82-4311-974c-ea3d5e58bf80" (UID: "f1d1d915-9b82-4311-974c-ea3d5e58bf80"). InnerVolumeSpecName "kube-api-access-55gwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.664328 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55gwc\" (UniqueName: \"kubernetes.io/projected/f1d1d915-9b82-4311-974c-ea3d5e58bf80-kube-api-access-55gwc\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.664367 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.664408 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.664421 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f1d1d915-9b82-4311-974c-ea3d5e58bf80-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.664431 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1d1d915-9b82-4311-974c-ea3d5e58bf80-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.665433 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.765239 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config\") pod \"5402a80f-2077-4207-81bc-cc12a2935886\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.765321 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-client-ca\") pod \"5402a80f-2077-4207-81bc-cc12a2935886\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.765443 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4vhc\" (UniqueName: \"kubernetes.io/projected/5402a80f-2077-4207-81bc-cc12a2935886-kube-api-access-p4vhc\") pod \"5402a80f-2077-4207-81bc-cc12a2935886\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.765479 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert\") pod \"5402a80f-2077-4207-81bc-cc12a2935886\" (UID: \"5402a80f-2077-4207-81bc-cc12a2935886\") " Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.766432 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-client-ca" (OuterVolumeSpecName: "client-ca") pod "5402a80f-2077-4207-81bc-cc12a2935886" (UID: "5402a80f-2077-4207-81bc-cc12a2935886"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.766535 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config" (OuterVolumeSpecName: "config") pod "5402a80f-2077-4207-81bc-cc12a2935886" (UID: "5402a80f-2077-4207-81bc-cc12a2935886"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.767248 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" event={"ID":"f1d1d915-9b82-4311-974c-ea3d5e58bf80","Type":"ContainerDied","Data":"23972bb0f2a4ae17cdc0ffcdcee4b14703555500354aff947101d609080337c0"} Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.767279 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8pc7r" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.767322 4744 scope.go:117] "RemoveContainer" containerID="b93b5e525d91494de8f362dd2451c35e648fce5a30674226d96f26ff925b6de0" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.770544 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5402a80f-2077-4207-81bc-cc12a2935886" (UID: "5402a80f-2077-4207-81bc-cc12a2935886"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.770975 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5402a80f-2077-4207-81bc-cc12a2935886-kube-api-access-p4vhc" (OuterVolumeSpecName: "kube-api-access-p4vhc") pod "5402a80f-2077-4207-81bc-cc12a2935886" (UID: "5402a80f-2077-4207-81bc-cc12a2935886"). InnerVolumeSpecName "kube-api-access-p4vhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.771248 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.771250 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9" event={"ID":"5402a80f-2077-4207-81bc-cc12a2935886","Type":"ContainerDied","Data":"a12a4514dd4bf050664ad0fd076c0f28d36ea039177c4930d7560123bfcfd24d"} Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.772375 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4dc8108a-af68-4f28-88eb-00bbb0ba134e","Type":"ContainerStarted","Data":"1a3dc7ccb02d57b14df9e7753550c7fe68f3c79842848d5f5db397f74d4a3d99"} Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.795250 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pc7r"] Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.799382 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8pc7r"] Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.833726 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9"] Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.839363 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-h7fn9"] Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.866545 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.866575 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5402a80f-2077-4207-81bc-cc12a2935886-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.866586 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4vhc\" (UniqueName: \"kubernetes.io/projected/5402a80f-2077-4207-81bc-cc12a2935886-kube-api-access-p4vhc\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:49 crc kubenswrapper[4744]: I0106 14:37:49.866596 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5402a80f-2077-4207-81bc-cc12a2935886-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.780023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4dc8108a-af68-4f28-88eb-00bbb0ba134e","Type":"ContainerStarted","Data":"ec765544e0aec46425de59e63d7982bf10aeb4c892b6e721ea3ddae3f1c084dd"} Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.800882 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=20.800864697 podStartE2EDuration="20.800864697s" podCreationTimestamp="2026-01-06 14:37:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:37:50.795387023 +0000 UTC m=+67.422853341" watchObservedRunningTime="2026-01-06 14:37:50.800864697 +0000 UTC m=+67.428331015" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.860289 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg"] Jan 06 14:37:50 crc kubenswrapper[4744]: E0106 14:37:50.860651 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0274d836-1692-4b10-b5f7-6282136d3ed6" containerName="pruner" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.860670 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0274d836-1692-4b10-b5f7-6282136d3ed6" containerName="pruner" Jan 06 14:37:50 crc kubenswrapper[4744]: E0106 14:37:50.860683 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5402a80f-2077-4207-81bc-cc12a2935886" containerName="route-controller-manager" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.860691 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5402a80f-2077-4207-81bc-cc12a2935886" containerName="route-controller-manager" Jan 06 14:37:50 crc kubenswrapper[4744]: E0106 14:37:50.860714 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1d1d915-9b82-4311-974c-ea3d5e58bf80" containerName="controller-manager" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.860724 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1d1d915-9b82-4311-974c-ea3d5e58bf80" containerName="controller-manager" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.860828 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1d1d915-9b82-4311-974c-ea3d5e58bf80" containerName="controller-manager" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.860839 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5402a80f-2077-4207-81bc-cc12a2935886" containerName="route-controller-manager" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.860847 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0274d836-1692-4b10-b5f7-6282136d3ed6" containerName="pruner" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.861330 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.863480 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.865621 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.865850 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.866041 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.866304 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.870652 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-65b4c59ffb-whpr6"] Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.870994 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.875601 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg"] Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.875734 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.878850 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thwk7\" (UniqueName: \"kubernetes.io/projected/b5f72773-ccea-4d9e-b7bc-613f76c1030f-kube-api-access-thwk7\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.878907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f72773-ccea-4d9e-b7bc-613f76c1030f-serving-cert\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.878959 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-config\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.879003 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-client-ca\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.884135 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-65b4c59ffb-whpr6"] Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.884734 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.884851 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.885045 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.885405 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.885860 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.886059 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.897023 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.980299 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-config\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.980343 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-proxy-ca-bundles\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.980380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-client-ca\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.980413 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thwk7\" (UniqueName: \"kubernetes.io/projected/b5f72773-ccea-4d9e-b7bc-613f76c1030f-kube-api-access-thwk7\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.980431 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-client-ca\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.980448 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhgq2\" (UniqueName: \"kubernetes.io/projected/9b17e310-1a30-46b3-b676-3f4fb0bae422-kube-api-access-hhgq2\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.980468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-config\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.980697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f72773-ccea-4d9e-b7bc-613f76c1030f-serving-cert\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:50 crc kubenswrapper[4744]: I0106 14:37:50.980721 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b17e310-1a30-46b3-b676-3f4fb0bae422-serving-cert\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.037592 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-client-ca\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.038515 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-config\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.054959 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f72773-ccea-4d9e-b7bc-613f76c1030f-serving-cert\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.061409 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thwk7\" (UniqueName: \"kubernetes.io/projected/b5f72773-ccea-4d9e-b7bc-613f76c1030f-kube-api-access-thwk7\") pod \"route-controller-manager-68c57f5dc6-4nmzg\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.081861 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-proxy-ca-bundles\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.081945 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-client-ca\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.082326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhgq2\" (UniqueName: \"kubernetes.io/projected/9b17e310-1a30-46b3-b676-3f4fb0bae422-kube-api-access-hhgq2\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.082356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-config\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.082392 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b17e310-1a30-46b3-b676-3f4fb0bae422-serving-cert\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.083472 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-client-ca\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.083625 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-proxy-ca-bundles\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.084823 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-config\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.085613 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b17e310-1a30-46b3-b676-3f4fb0bae422-serving-cert\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.097806 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhgq2\" (UniqueName: \"kubernetes.io/projected/9b17e310-1a30-46b3-b676-3f4fb0bae422-kube-api-access-hhgq2\") pod \"controller-manager-65b4c59ffb-whpr6\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.210621 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.338461 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.726526 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5402a80f-2077-4207-81bc-cc12a2935886" path="/var/lib/kubelet/pods/5402a80f-2077-4207-81bc-cc12a2935886/volumes" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.727211 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1d1d915-9b82-4311-974c-ea3d5e58bf80" path="/var/lib/kubelet/pods/f1d1d915-9b82-4311-974c-ea3d5e58bf80/volumes" Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.790118 4744 generic.go:334] "Generic (PLEG): container finished" podID="4dc8108a-af68-4f28-88eb-00bbb0ba134e" containerID="ec765544e0aec46425de59e63d7982bf10aeb4c892b6e721ea3ddae3f1c084dd" exitCode=0 Jan 06 14:37:51 crc kubenswrapper[4744]: I0106 14:37:51.790219 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4dc8108a-af68-4f28-88eb-00bbb0ba134e","Type":"ContainerDied","Data":"ec765544e0aec46425de59e63d7982bf10aeb4c892b6e721ea3ddae3f1c084dd"} Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.778776 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.824744 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4dc8108a-af68-4f28-88eb-00bbb0ba134e","Type":"ContainerDied","Data":"1a3dc7ccb02d57b14df9e7753550c7fe68f3c79842848d5f5db397f74d4a3d99"} Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.825194 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a3dc7ccb02d57b14df9e7753550c7fe68f3c79842848d5f5db397f74d4a3d99" Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.824760 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.827835 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-bd9mm_df8c0d6b-5dc0-42cc-90c0-c254f463275b/kube-multus-additional-cni-plugins/0.log" Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.827895 4744 generic.go:334] "Generic (PLEG): container finished" podID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" exitCode=137 Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.827930 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" event={"ID":"df8c0d6b-5dc0-42cc-90c0-c254f463275b","Type":"ContainerDied","Data":"cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db"} Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.891951 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kube-api-access\") pod \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\" (UID: \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\") " Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.892009 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kubelet-dir\") pod \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\" (UID: \"4dc8108a-af68-4f28-88eb-00bbb0ba134e\") " Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.892144 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4dc8108a-af68-4f28-88eb-00bbb0ba134e" (UID: "4dc8108a-af68-4f28-88eb-00bbb0ba134e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.900513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4dc8108a-af68-4f28-88eb-00bbb0ba134e" (UID: "4dc8108a-af68-4f28-88eb-00bbb0ba134e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.993946 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:55 crc kubenswrapper[4744]: I0106 14:37:55.993984 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4dc8108a-af68-4f28-88eb-00bbb0ba134e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:37:58 crc kubenswrapper[4744]: I0106 14:37:58.129297 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" Jan 06 14:37:58 crc kubenswrapper[4744]: E0106 14:37:58.265540 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db is running failed: container process not found" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:58 crc kubenswrapper[4744]: E0106 14:37:58.266184 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db is running failed: container process not found" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:58 crc kubenswrapper[4744]: E0106 14:37:58.266580 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db is running failed: container process not found" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:37:58 crc kubenswrapper[4744]: E0106 14:37:58.266611 4744 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerName="kube-multus-additional-cni-plugins" Jan 06 14:37:59 crc kubenswrapper[4744]: I0106 14:37:59.328694 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 06 14:38:00 crc kubenswrapper[4744]: I0106 14:38:00.907727 4744 scope.go:117] "RemoveContainer" containerID="229a0808bcdf0ebfd4f50e30b9f4ab9973979e575ad9fa32f22baea299c3b9e4" Jan 06 14:38:02 crc kubenswrapper[4744]: I0106 14:38:02.726930 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 06 14:38:03 crc kubenswrapper[4744]: I0106 14:38:03.201284 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65b4c59ffb-whpr6"] Jan 06 14:38:03 crc kubenswrapper[4744]: I0106 14:38:03.225966 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg"] Jan 06 14:38:03 crc kubenswrapper[4744]: I0106 14:38:03.728184 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.728138401 podStartE2EDuration="1.728138401s" podCreationTimestamp="2026-01-06 14:38:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:38:03.725538959 +0000 UTC m=+80.353005277" watchObservedRunningTime="2026-01-06 14:38:03.728138401 +0000 UTC m=+80.355604729" Jan 06 14:38:03 crc kubenswrapper[4744]: E0106 14:38:03.734957 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 06 14:38:03 crc kubenswrapper[4744]: E0106 14:38:03.735202 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-49v7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-m4xw2_openshift-marketplace(1d41f99a-8352-472d-9011-e9c0259abf7b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 06 14:38:03 crc kubenswrapper[4744]: E0106 14:38:03.736423 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-m4xw2" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" Jan 06 14:38:07 crc kubenswrapper[4744]: E0106 14:38:07.019980 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-m4xw2" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" Jan 06 14:38:07 crc kubenswrapper[4744]: E0106 14:38:07.087574 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 06 14:38:07 crc kubenswrapper[4744]: E0106 14:38:07.087724 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jhbt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-69twj_openshift-marketplace(6cbba207-bda4-4f95-a8f1-d003787cb231): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 06 14:38:07 crc kubenswrapper[4744]: E0106 14:38:07.088908 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-69twj" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" Jan 06 14:38:07 crc kubenswrapper[4744]: E0106 14:38:07.128124 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 06 14:38:07 crc kubenswrapper[4744]: E0106 14:38:07.128283 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-68bnt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4bh4q_openshift-marketplace(408386a1-c1c3-4b95-8bce-4c60433159a0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 06 14:38:07 crc kubenswrapper[4744]: E0106 14:38:07.129503 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-4bh4q" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.181613 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-4bh4q" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.181877 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-69twj" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.266477 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.266834 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db is running failed: container process not found" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.266935 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jh2lp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4hxcm_openshift-marketplace(6fe6e90c-e807-441c-b130-15a7368b87f0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.267943 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db is running failed: container process not found" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.268083 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4hxcm" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.268347 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db is running failed: container process not found" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" cmd=["/bin/bash","-c","test -f /ready/ready"] Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.268407 4744 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerName="kube-multus-additional-cni-plugins" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.275512 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.275631 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wcvmd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-fv9qm_openshift-marketplace(b1d9691a-454d-4ec0-9447-5cf0df1f5cfe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.277357 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-fv9qm" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.286991 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-bd9mm_df8c0d6b-5dc0-42cc-90c0-c254f463275b/kube-multus-additional-cni-plugins/0.log" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.287074 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.309412 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.309544 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-98h5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4cwvb_openshift-marketplace(14d7cfd7-181f-4d79-8951-27d300d49a8f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.310751 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4cwvb" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.384629 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/df8c0d6b-5dc0-42cc-90c0-c254f463275b-ready\") pod \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.384975 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/df8c0d6b-5dc0-42cc-90c0-c254f463275b-cni-sysctl-allowlist\") pod \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.384999 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/df8c0d6b-5dc0-42cc-90c0-c254f463275b-tuning-conf-dir\") pod \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.385047 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df8c0d6b-5dc0-42cc-90c0-c254f463275b-ready" (OuterVolumeSpecName: "ready") pod "df8c0d6b-5dc0-42cc-90c0-c254f463275b" (UID: "df8c0d6b-5dc0-42cc-90c0-c254f463275b"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.385074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn5fp\" (UniqueName: \"kubernetes.io/projected/df8c0d6b-5dc0-42cc-90c0-c254f463275b-kube-api-access-cn5fp\") pod \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\" (UID: \"df8c0d6b-5dc0-42cc-90c0-c254f463275b\") " Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.385096 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df8c0d6b-5dc0-42cc-90c0-c254f463275b-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "df8c0d6b-5dc0-42cc-90c0-c254f463275b" (UID: "df8c0d6b-5dc0-42cc-90c0-c254f463275b"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.385661 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df8c0d6b-5dc0-42cc-90c0-c254f463275b-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "df8c0d6b-5dc0-42cc-90c0-c254f463275b" (UID: "df8c0d6b-5dc0-42cc-90c0-c254f463275b"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.386531 4744 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/df8c0d6b-5dc0-42cc-90c0-c254f463275b-ready\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.386549 4744 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/df8c0d6b-5dc0-42cc-90c0-c254f463275b-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.386560 4744 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/df8c0d6b-5dc0-42cc-90c0-c254f463275b-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.390572 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df8c0d6b-5dc0-42cc-90c0-c254f463275b-kube-api-access-cn5fp" (OuterVolumeSpecName: "kube-api-access-cn5fp") pod "df8c0d6b-5dc0-42cc-90c0-c254f463275b" (UID: "df8c0d6b-5dc0-42cc-90c0-c254f463275b"). InnerVolumeSpecName "kube-api-access-cn5fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.440077 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg"] Jan 06 14:38:08 crc kubenswrapper[4744]: W0106 14:38:08.450803 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5f72773_ccea_4d9e_b7bc_613f76c1030f.slice/crio-e1b1e5b80634f3d61f009a6de8766eedc6028c09c71e18ffdda3526869947df9 WatchSource:0}: Error finding container e1b1e5b80634f3d61f009a6de8766eedc6028c09c71e18ffdda3526869947df9: Status 404 returned error can't find the container with id e1b1e5b80634f3d61f009a6de8766eedc6028c09c71e18ffdda3526869947df9 Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.489418 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn5fp\" (UniqueName: \"kubernetes.io/projected/df8c0d6b-5dc0-42cc-90c0-c254f463275b-kube-api-access-cn5fp\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.621758 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65b4c59ffb-whpr6"] Jan 06 14:38:08 crc kubenswrapper[4744]: W0106 14:38:08.622737 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b17e310_1a30_46b3_b676_3f4fb0bae422.slice/crio-c9e0094a3a0702567a9a1d35d470bcf9c3490d9fbe4269968f721ee9860ea6b5 WatchSource:0}: Error finding container c9e0094a3a0702567a9a1d35d470bcf9c3490d9fbe4269968f721ee9860ea6b5: Status 404 returned error can't find the container with id c9e0094a3a0702567a9a1d35d470bcf9c3490d9fbe4269968f721ee9860ea6b5 Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.905188 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-bd9mm_df8c0d6b-5dc0-42cc-90c0-c254f463275b/kube-multus-additional-cni-plugins/0.log" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.905505 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.906414 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-bd9mm" event={"ID":"df8c0d6b-5dc0-42cc-90c0-c254f463275b","Type":"ContainerDied","Data":"22b032a283dd099d38837b08a53cc9e5be13ebafb82f46c058dba83be9de7fef"} Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.906467 4744 scope.go:117] "RemoveContainer" containerID="cca9dad39eeae85f36e6d49261cd5a329deecda2aa8b04add323b469312009db" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.907980 4744 generic.go:334] "Generic (PLEG): container finished" podID="ead06eb7-e94b-400a-b568-1871fea3e807" containerID="d4677c4cb280809b533e20341865dcd6a84861cade2772820f6882c5a3665550" exitCode=0 Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.908043 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8458" event={"ID":"ead06eb7-e94b-400a-b568-1871fea3e807","Type":"ContainerDied","Data":"d4677c4cb280809b533e20341865dcd6a84861cade2772820f6882c5a3665550"} Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.914649 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" event={"ID":"b5f72773-ccea-4d9e-b7bc-613f76c1030f","Type":"ContainerStarted","Data":"90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0"} Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.914694 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" event={"ID":"b5f72773-ccea-4d9e-b7bc-613f76c1030f","Type":"ContainerStarted","Data":"e1b1e5b80634f3d61f009a6de8766eedc6028c09c71e18ffdda3526869947df9"} Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.914792 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" podUID="b5f72773-ccea-4d9e-b7bc-613f76c1030f" containerName="route-controller-manager" containerID="cri-o://90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0" gracePeriod=30 Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.915040 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.917799 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" event={"ID":"9b17e310-1a30-46b3-b676-3f4fb0bae422","Type":"ContainerStarted","Data":"c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24"} Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.917835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" event={"ID":"9b17e310-1a30-46b3-b676-3f4fb0bae422","Type":"ContainerStarted","Data":"c9e0094a3a0702567a9a1d35d470bcf9c3490d9fbe4269968f721ee9860ea6b5"} Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.917943 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" podUID="9b17e310-1a30-46b3-b676-3f4fb0bae422" containerName="controller-manager" containerID="cri-o://c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24" gracePeriod=30 Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.918253 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.919341 4744 patch_prober.go:28] interesting pod/controller-manager-65b4c59ffb-whpr6 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" start-of-body= Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.919367 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" podUID="9b17e310-1a30-46b3-b676-3f4fb0bae422" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.923281 4744 generic.go:334] "Generic (PLEG): container finished" podID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerID="ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665" exitCode=0 Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.923917 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j999h" event={"ID":"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf","Type":"ContainerDied","Data":"ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665"} Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.927506 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-fv9qm" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.927881 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4hxcm" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" Jan 06 14:38:08 crc kubenswrapper[4744]: E0106 14:38:08.931418 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4cwvb" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.957624 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-bd9mm"] Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.960597 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-bd9mm"] Jan 06 14:38:08 crc kubenswrapper[4744]: I0106 14:38:08.996473 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" podStartSLOduration=25.996457949 podStartE2EDuration="25.996457949s" podCreationTimestamp="2026-01-06 14:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:38:08.994961148 +0000 UTC m=+85.622427476" watchObservedRunningTime="2026-01-06 14:38:08.996457949 +0000 UTC m=+85.623924267" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.034674 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" podStartSLOduration=26.034655212 podStartE2EDuration="26.034655212s" podCreationTimestamp="2026-01-06 14:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:38:09.031534475 +0000 UTC m=+85.659000793" watchObservedRunningTime="2026-01-06 14:38:09.034655212 +0000 UTC m=+85.662121550" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.281334 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-65b4c59ffb-whpr6_9b17e310-1a30-46b3-b676-3f4fb0bae422/controller-manager/0.log" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.281411 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.399330 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhgq2\" (UniqueName: \"kubernetes.io/projected/9b17e310-1a30-46b3-b676-3f4fb0bae422-kube-api-access-hhgq2\") pod \"9b17e310-1a30-46b3-b676-3f4fb0bae422\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.399431 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b17e310-1a30-46b3-b676-3f4fb0bae422-serving-cert\") pod \"9b17e310-1a30-46b3-b676-3f4fb0bae422\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.399453 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-proxy-ca-bundles\") pod \"9b17e310-1a30-46b3-b676-3f4fb0bae422\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.399482 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-client-ca\") pod \"9b17e310-1a30-46b3-b676-3f4fb0bae422\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.399525 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-config\") pod \"9b17e310-1a30-46b3-b676-3f4fb0bae422\" (UID: \"9b17e310-1a30-46b3-b676-3f4fb0bae422\") " Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.400612 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-client-ca" (OuterVolumeSpecName: "client-ca") pod "9b17e310-1a30-46b3-b676-3f4fb0bae422" (UID: "9b17e310-1a30-46b3-b676-3f4fb0bae422"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.400991 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9b17e310-1a30-46b3-b676-3f4fb0bae422" (UID: "9b17e310-1a30-46b3-b676-3f4fb0bae422"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.401733 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-config" (OuterVolumeSpecName: "config") pod "9b17e310-1a30-46b3-b676-3f4fb0bae422" (UID: "9b17e310-1a30-46b3-b676-3f4fb0bae422"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.405587 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b17e310-1a30-46b3-b676-3f4fb0bae422-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9b17e310-1a30-46b3-b676-3f4fb0bae422" (UID: "9b17e310-1a30-46b3-b676-3f4fb0bae422"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.407188 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b17e310-1a30-46b3-b676-3f4fb0bae422-kube-api-access-hhgq2" (OuterVolumeSpecName: "kube-api-access-hhgq2") pod "9b17e310-1a30-46b3-b676-3f4fb0bae422" (UID: "9b17e310-1a30-46b3-b676-3f4fb0bae422"). InnerVolumeSpecName "kube-api-access-hhgq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.418290 4744 patch_prober.go:28] interesting pod/route-controller-manager-68c57f5dc6-4nmzg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": read tcp 10.217.0.2:41144->10.217.0.54:8443: read: connection reset by peer" start-of-body= Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.418375 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" podUID="b5f72773-ccea-4d9e-b7bc-613f76c1030f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": read tcp 10.217.0.2:41144->10.217.0.54:8443: read: connection reset by peer" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.501334 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhgq2\" (UniqueName: \"kubernetes.io/projected/9b17e310-1a30-46b3-b676-3f4fb0bae422-kube-api-access-hhgq2\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.501502 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b17e310-1a30-46b3-b676-3f4fb0bae422-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.501516 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.501527 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.501539 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b17e310-1a30-46b3-b676-3f4fb0bae422-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.669012 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-68c57f5dc6-4nmzg_b5f72773-ccea-4d9e-b7bc-613f76c1030f/route-controller-manager/0.log" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.669155 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.718992 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" path="/var/lib/kubelet/pods/df8c0d6b-5dc0-42cc-90c0-c254f463275b/volumes" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.806280 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-config\") pod \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.806334 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f72773-ccea-4d9e-b7bc-613f76c1030f-serving-cert\") pod \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.806370 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-client-ca\") pod \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.806414 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thwk7\" (UniqueName: \"kubernetes.io/projected/b5f72773-ccea-4d9e-b7bc-613f76c1030f-kube-api-access-thwk7\") pod \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\" (UID: \"b5f72773-ccea-4d9e-b7bc-613f76c1030f\") " Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.807048 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-config" (OuterVolumeSpecName: "config") pod "b5f72773-ccea-4d9e-b7bc-613f76c1030f" (UID: "b5f72773-ccea-4d9e-b7bc-613f76c1030f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.807142 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-client-ca" (OuterVolumeSpecName: "client-ca") pod "b5f72773-ccea-4d9e-b7bc-613f76c1030f" (UID: "b5f72773-ccea-4d9e-b7bc-613f76c1030f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.810855 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5f72773-ccea-4d9e-b7bc-613f76c1030f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b5f72773-ccea-4d9e-b7bc-613f76c1030f" (UID: "b5f72773-ccea-4d9e-b7bc-613f76c1030f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.811870 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5f72773-ccea-4d9e-b7bc-613f76c1030f-kube-api-access-thwk7" (OuterVolumeSpecName: "kube-api-access-thwk7") pod "b5f72773-ccea-4d9e-b7bc-613f76c1030f" (UID: "b5f72773-ccea-4d9e-b7bc-613f76c1030f"). InnerVolumeSpecName "kube-api-access-thwk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.910655 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5f72773-ccea-4d9e-b7bc-613f76c1030f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.910700 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.910711 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thwk7\" (UniqueName: \"kubernetes.io/projected/b5f72773-ccea-4d9e-b7bc-613f76c1030f-kube-api-access-thwk7\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.910721 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5f72773-ccea-4d9e-b7bc-613f76c1030f-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.933185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j999h" event={"ID":"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf","Type":"ContainerStarted","Data":"ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d"} Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.936896 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8458" event={"ID":"ead06eb7-e94b-400a-b568-1871fea3e807","Type":"ContainerStarted","Data":"9f828e570440772fe4175a793636c299683646545d0de22a97fdd931f3ac0a14"} Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.938882 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-68c57f5dc6-4nmzg_b5f72773-ccea-4d9e-b7bc-613f76c1030f/route-controller-manager/0.log" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.938946 4744 generic.go:334] "Generic (PLEG): container finished" podID="b5f72773-ccea-4d9e-b7bc-613f76c1030f" containerID="90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0" exitCode=255 Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.939002 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.939071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" event={"ID":"b5f72773-ccea-4d9e-b7bc-613f76c1030f","Type":"ContainerDied","Data":"90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0"} Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.939117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg" event={"ID":"b5f72773-ccea-4d9e-b7bc-613f76c1030f","Type":"ContainerDied","Data":"e1b1e5b80634f3d61f009a6de8766eedc6028c09c71e18ffdda3526869947df9"} Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.939143 4744 scope.go:117] "RemoveContainer" containerID="90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.940859 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-65b4c59ffb-whpr6_9b17e310-1a30-46b3-b676-3f4fb0bae422/controller-manager/0.log" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.940933 4744 generic.go:334] "Generic (PLEG): container finished" podID="9b17e310-1a30-46b3-b676-3f4fb0bae422" containerID="c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24" exitCode=2 Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.940984 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.940980 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" event={"ID":"9b17e310-1a30-46b3-b676-3f4fb0bae422","Type":"ContainerDied","Data":"c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24"} Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.941149 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b4c59ffb-whpr6" event={"ID":"9b17e310-1a30-46b3-b676-3f4fb0bae422","Type":"ContainerDied","Data":"c9e0094a3a0702567a9a1d35d470bcf9c3490d9fbe4269968f721ee9860ea6b5"} Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.953378 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j999h" podStartSLOduration=4.04946732 podStartE2EDuration="49.953359067s" podCreationTimestamp="2026-01-06 14:37:20 +0000 UTC" firstStartedPulling="2026-01-06 14:37:23.528957146 +0000 UTC m=+40.156423464" lastFinishedPulling="2026-01-06 14:38:09.432848893 +0000 UTC m=+86.060315211" observedRunningTime="2026-01-06 14:38:09.950315622 +0000 UTC m=+86.577781980" watchObservedRunningTime="2026-01-06 14:38:09.953359067 +0000 UTC m=+86.580825385" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.963401 4744 scope.go:117] "RemoveContainer" containerID="90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0" Jan 06 14:38:09 crc kubenswrapper[4744]: E0106 14:38:09.963858 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0\": container with ID starting with 90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0 not found: ID does not exist" containerID="90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.963996 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0"} err="failed to get container status \"90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0\": rpc error: code = NotFound desc = could not find container \"90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0\": container with ID starting with 90edd15d2d6686c5e81609188a220c5e03c06dc259f6e5e7a7c02497875594b0 not found: ID does not exist" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.964108 4744 scope.go:117] "RemoveContainer" containerID="c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.980259 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 06 14:38:09 crc kubenswrapper[4744]: E0106 14:38:09.980495 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b17e310-1a30-46b3-b676-3f4fb0bae422" containerName="controller-manager" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.980515 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b17e310-1a30-46b3-b676-3f4fb0bae422" containerName="controller-manager" Jan 06 14:38:09 crc kubenswrapper[4744]: E0106 14:38:09.980532 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc8108a-af68-4f28-88eb-00bbb0ba134e" containerName="pruner" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.980541 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc8108a-af68-4f28-88eb-00bbb0ba134e" containerName="pruner" Jan 06 14:38:09 crc kubenswrapper[4744]: E0106 14:38:09.980561 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5f72773-ccea-4d9e-b7bc-613f76c1030f" containerName="route-controller-manager" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.980570 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5f72773-ccea-4d9e-b7bc-613f76c1030f" containerName="route-controller-manager" Jan 06 14:38:09 crc kubenswrapper[4744]: E0106 14:38:09.980584 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerName="kube-multus-additional-cni-plugins" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.980594 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerName="kube-multus-additional-cni-plugins" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.980712 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dc8108a-af68-4f28-88eb-00bbb0ba134e" containerName="pruner" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.980726 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b17e310-1a30-46b3-b676-3f4fb0bae422" containerName="controller-manager" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.980740 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="df8c0d6b-5dc0-42cc-90c0-c254f463275b" containerName="kube-multus-additional-cni-plugins" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.980750 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5f72773-ccea-4d9e-b7bc-613f76c1030f" containerName="route-controller-manager" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.981140 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.981929 4744 scope.go:117] "RemoveContainer" containerID="c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.982643 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c8458" podStartSLOduration=4.283218583 podStartE2EDuration="49.982621741s" podCreationTimestamp="2026-01-06 14:37:20 +0000 UTC" firstStartedPulling="2026-01-06 14:37:23.745201592 +0000 UTC m=+40.372667910" lastFinishedPulling="2026-01-06 14:38:09.44460475 +0000 UTC m=+86.072071068" observedRunningTime="2026-01-06 14:38:09.9797087 +0000 UTC m=+86.607175028" watchObservedRunningTime="2026-01-06 14:38:09.982621741 +0000 UTC m=+86.610088059" Jan 06 14:38:09 crc kubenswrapper[4744]: E0106 14:38:09.984257 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24\": container with ID starting with c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24 not found: ID does not exist" containerID="c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.984385 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24"} err="failed to get container status \"c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24\": rpc error: code = NotFound desc = could not find container \"c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24\": container with ID starting with c7438991b4bb9c9cc946225ef4a846bef428827156ffbf50614ba517722d4f24 not found: ID does not exist" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.984492 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.984680 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 06 14:38:09 crc kubenswrapper[4744]: I0106 14:38:09.993243 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.006746 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65b4c59ffb-whpr6"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.008944 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-65b4c59ffb-whpr6"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.012369 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51c58bb1-a89f-4f25-9c02-ee954db9f178-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51c58bb1-a89f-4f25-9c02-ee954db9f178\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.012452 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51c58bb1-a89f-4f25-9c02-ee954db9f178-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51c58bb1-a89f-4f25-9c02-ee954db9f178\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.032489 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.034672 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-68c57f5dc6-4nmzg"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.114136 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51c58bb1-a89f-4f25-9c02-ee954db9f178-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51c58bb1-a89f-4f25-9c02-ee954db9f178\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.114253 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51c58bb1-a89f-4f25-9c02-ee954db9f178-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"51c58bb1-a89f-4f25-9c02-ee954db9f178\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.114259 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51c58bb1-a89f-4f25-9c02-ee954db9f178-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51c58bb1-a89f-4f25-9c02-ee954db9f178\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.132014 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51c58bb1-a89f-4f25-9c02-ee954db9f178-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"51c58bb1-a89f-4f25-9c02-ee954db9f178\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.307802 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.556966 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.895063 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-86f674b85d-l6b67"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.896094 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.900648 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.900804 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.900934 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.901061 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.901188 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.902095 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.903301 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.904185 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.908087 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.908340 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.908517 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.908590 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.908658 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.909632 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.913329 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.914081 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86f674b85d-l6b67"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.921016 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5"] Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.923268 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldk9l\" (UniqueName: \"kubernetes.io/projected/7273d766-2526-4fbe-836a-3af7c3eb37cf-kube-api-access-ldk9l\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.923565 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-config\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.923769 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-config\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.924206 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdd9t\" (UniqueName: \"kubernetes.io/projected/5f926310-2075-4937-b88e-b6fd6184ef70-kube-api-access-wdd9t\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.924419 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7273d766-2526-4fbe-836a-3af7c3eb37cf-serving-cert\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.924584 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-proxy-ca-bundles\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.924704 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-client-ca\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.924935 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-client-ca\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.925184 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f926310-2075-4937-b88e-b6fd6184ef70-serving-cert\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.951768 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"51c58bb1-a89f-4f25-9c02-ee954db9f178","Type":"ContainerStarted","Data":"a325ad0c6af660b45d50ea58c34a69458ab88be5dec000d87102b5bebaf0ad13"} Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.951809 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"51c58bb1-a89f-4f25-9c02-ee954db9f178","Type":"ContainerStarted","Data":"b540d4e7abb2be43ccfbf74fb495731acb887bc0a7569b29be7c7b22e2215bbe"} Jan 06 14:38:10 crc kubenswrapper[4744]: I0106 14:38:10.970469 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.970452518 podStartE2EDuration="1.970452518s" podCreationTimestamp="2026-01-06 14:38:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:38:10.967113925 +0000 UTC m=+87.594580253" watchObservedRunningTime="2026-01-06 14:38:10.970452518 +0000 UTC m=+87.597918836" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.026592 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-client-ca\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.026659 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f926310-2075-4937-b88e-b6fd6184ef70-serving-cert\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.026708 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldk9l\" (UniqueName: \"kubernetes.io/projected/7273d766-2526-4fbe-836a-3af7c3eb37cf-kube-api-access-ldk9l\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.026781 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-config\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.026802 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-config\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.026824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdd9t\" (UniqueName: \"kubernetes.io/projected/5f926310-2075-4937-b88e-b6fd6184ef70-kube-api-access-wdd9t\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.026845 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7273d766-2526-4fbe-836a-3af7c3eb37cf-serving-cert\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.026868 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-proxy-ca-bundles\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.026904 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-client-ca\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.028245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-config\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.029219 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-config\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.028255 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-client-ca\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.029992 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-client-ca\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.030974 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-proxy-ca-bundles\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.034537 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7273d766-2526-4fbe-836a-3af7c3eb37cf-serving-cert\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.036760 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f926310-2075-4937-b88e-b6fd6184ef70-serving-cert\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.039169 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.039278 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.048013 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdd9t\" (UniqueName: \"kubernetes.io/projected/5f926310-2075-4937-b88e-b6fd6184ef70-kube-api-access-wdd9t\") pod \"controller-manager-86f674b85d-l6b67\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.053387 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldk9l\" (UniqueName: \"kubernetes.io/projected/7273d766-2526-4fbe-836a-3af7c3eb37cf-kube-api-access-ldk9l\") pod \"route-controller-manager-796798f6b5-jm7x5\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.214724 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.221590 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.393542 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.394293 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.446017 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86f674b85d-l6b67"] Jan 06 14:38:11 crc kubenswrapper[4744]: W0106 14:38:11.459341 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f926310_2075_4937_b88e_b6fd6184ef70.slice/crio-b9bdccf2cbf585eacebdf034a8959afee01236284eaaca4482183a808127d58c WatchSource:0}: Error finding container b9bdccf2cbf585eacebdf034a8959afee01236284eaaca4482183a808127d58c: Status 404 returned error can't find the container with id b9bdccf2cbf585eacebdf034a8959afee01236284eaaca4482183a808127d58c Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.477022 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5"] Jan 06 14:38:11 crc kubenswrapper[4744]: W0106 14:38:11.486459 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7273d766_2526_4fbe_836a_3af7c3eb37cf.slice/crio-8920aba6571811643b97da00779d64b7212b684cd06d16eb9dca3594e2bc263f WatchSource:0}: Error finding container 8920aba6571811643b97da00779d64b7212b684cd06d16eb9dca3594e2bc263f: Status 404 returned error can't find the container with id 8920aba6571811643b97da00779d64b7212b684cd06d16eb9dca3594e2bc263f Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.718065 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b17e310-1a30-46b3-b676-3f4fb0bae422" path="/var/lib/kubelet/pods/9b17e310-1a30-46b3-b676-3f4fb0bae422/volumes" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.718999 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5f72773-ccea-4d9e-b7bc-613f76c1030f" path="/var/lib/kubelet/pods/b5f72773-ccea-4d9e-b7bc-613f76c1030f/volumes" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.960801 4744 generic.go:334] "Generic (PLEG): container finished" podID="51c58bb1-a89f-4f25-9c02-ee954db9f178" containerID="a325ad0c6af660b45d50ea58c34a69458ab88be5dec000d87102b5bebaf0ad13" exitCode=0 Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.960912 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"51c58bb1-a89f-4f25-9c02-ee954db9f178","Type":"ContainerDied","Data":"a325ad0c6af660b45d50ea58c34a69458ab88be5dec000d87102b5bebaf0ad13"} Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.963633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" event={"ID":"5f926310-2075-4937-b88e-b6fd6184ef70","Type":"ContainerStarted","Data":"7438288f53f0a5ac41617027d4316717a98e0e64d9053d58749cf5b2b9bdf0b2"} Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.966216 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.966273 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.966293 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" event={"ID":"5f926310-2075-4937-b88e-b6fd6184ef70","Type":"ContainerStarted","Data":"b9bdccf2cbf585eacebdf034a8959afee01236284eaaca4482183a808127d58c"} Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.966324 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" event={"ID":"7273d766-2526-4fbe-836a-3af7c3eb37cf","Type":"ContainerStarted","Data":"bfd235c30e5fb7cf3a5af31593e5d836a8220dd40b1c88a86e297e8246adb6c2"} Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.966339 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" event={"ID":"7273d766-2526-4fbe-836a-3af7c3eb37cf","Type":"ContainerStarted","Data":"8920aba6571811643b97da00779d64b7212b684cd06d16eb9dca3594e2bc263f"} Jan 06 14:38:11 crc kubenswrapper[4744]: I0106 14:38:11.968082 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:12 crc kubenswrapper[4744]: I0106 14:38:12.005994 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" podStartSLOduration=9.005976083 podStartE2EDuration="9.005976083s" podCreationTimestamp="2026-01-06 14:38:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:38:12.003645468 +0000 UTC m=+88.631111786" watchObservedRunningTime="2026-01-06 14:38:12.005976083 +0000 UTC m=+88.633442401" Jan 06 14:38:12 crc kubenswrapper[4744]: I0106 14:38:12.107134 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-c8458" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" containerName="registry-server" probeResult="failure" output=< Jan 06 14:38:12 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:38:12 crc kubenswrapper[4744]: > Jan 06 14:38:12 crc kubenswrapper[4744]: I0106 14:38:12.280478 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:12 crc kubenswrapper[4744]: I0106 14:38:12.307855 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" podStartSLOduration=9.307833743 podStartE2EDuration="9.307833743s" podCreationTimestamp="2026-01-06 14:38:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:38:12.030993939 +0000 UTC m=+88.658460277" watchObservedRunningTime="2026-01-06 14:38:12.307833743 +0000 UTC m=+88.935300061" Jan 06 14:38:12 crc kubenswrapper[4744]: I0106 14:38:12.438010 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-j999h" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerName="registry-server" probeResult="failure" output=< Jan 06 14:38:12 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:38:12 crc kubenswrapper[4744]: > Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.290925 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.475227 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51c58bb1-a89f-4f25-9c02-ee954db9f178-kubelet-dir\") pod \"51c58bb1-a89f-4f25-9c02-ee954db9f178\" (UID: \"51c58bb1-a89f-4f25-9c02-ee954db9f178\") " Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.475311 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/51c58bb1-a89f-4f25-9c02-ee954db9f178-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "51c58bb1-a89f-4f25-9c02-ee954db9f178" (UID: "51c58bb1-a89f-4f25-9c02-ee954db9f178"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.475331 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51c58bb1-a89f-4f25-9c02-ee954db9f178-kube-api-access\") pod \"51c58bb1-a89f-4f25-9c02-ee954db9f178\" (UID: \"51c58bb1-a89f-4f25-9c02-ee954db9f178\") " Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.475777 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/51c58bb1-a89f-4f25-9c02-ee954db9f178-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.497656 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51c58bb1-a89f-4f25-9c02-ee954db9f178-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "51c58bb1-a89f-4f25-9c02-ee954db9f178" (UID: "51c58bb1-a89f-4f25-9c02-ee954db9f178"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.576723 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/51c58bb1-a89f-4f25-9c02-ee954db9f178-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.978859 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"51c58bb1-a89f-4f25-9c02-ee954db9f178","Type":"ContainerDied","Data":"b540d4e7abb2be43ccfbf74fb495731acb887bc0a7569b29be7c7b22e2215bbe"} Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.978918 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b540d4e7abb2be43ccfbf74fb495731acb887bc0a7569b29be7c7b22e2215bbe" Jan 06 14:38:13 crc kubenswrapper[4744]: I0106 14:38:13.978962 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.789355 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 06 14:38:16 crc kubenswrapper[4744]: E0106 14:38:16.790036 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51c58bb1-a89f-4f25-9c02-ee954db9f178" containerName="pruner" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.790058 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="51c58bb1-a89f-4f25-9c02-ee954db9f178" containerName="pruner" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.790299 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="51c58bb1-a89f-4f25-9c02-ee954db9f178" containerName="pruner" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.790859 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.799935 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.800276 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.820344 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.820408 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7684584-6c37-4d30-8593-70d738b41d93-kube-api-access\") pod \"installer-9-crc\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.820464 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-var-lock\") pod \"installer-9-crc\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.821028 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.921277 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7684584-6c37-4d30-8593-70d738b41d93-kube-api-access\") pod \"installer-9-crc\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.921354 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-var-lock\") pod \"installer-9-crc\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.921441 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.921509 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.921549 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-var-lock\") pod \"installer-9-crc\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:16 crc kubenswrapper[4744]: I0106 14:38:16.952604 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7684584-6c37-4d30-8593-70d738b41d93-kube-api-access\") pod \"installer-9-crc\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:17 crc kubenswrapper[4744]: I0106 14:38:17.116113 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:17 crc kubenswrapper[4744]: I0106 14:38:17.608510 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 06 14:38:18 crc kubenswrapper[4744]: I0106 14:38:18.001144 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c7684584-6c37-4d30-8593-70d738b41d93","Type":"ContainerStarted","Data":"04b6ad349bf54a1fd227cccd861666308d911b9fe43864e803c9ee5e831dbf7d"} Jan 06 14:38:20 crc kubenswrapper[4744]: I0106 14:38:20.016683 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c7684584-6c37-4d30-8593-70d738b41d93","Type":"ContainerStarted","Data":"f3841de24ff12e431a5c37adf0bbe9b220965fc44b868a99ef4964b9c58de9cf"} Jan 06 14:38:20 crc kubenswrapper[4744]: I0106 14:38:20.019293 4744 generic.go:334] "Generic (PLEG): container finished" podID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerID="c1b039cd6c94530f4507938fcb9cd6902a45038c22828cd02ec0df9dc77c190f" exitCode=0 Jan 06 14:38:20 crc kubenswrapper[4744]: I0106 14:38:20.019338 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4xw2" event={"ID":"1d41f99a-8352-472d-9011-e9c0259abf7b","Type":"ContainerDied","Data":"c1b039cd6c94530f4507938fcb9cd6902a45038c22828cd02ec0df9dc77c190f"} Jan 06 14:38:20 crc kubenswrapper[4744]: I0106 14:38:20.029718 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=4.029694254 podStartE2EDuration="4.029694254s" podCreationTimestamp="2026-01-06 14:38:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:38:20.029259842 +0000 UTC m=+96.656726160" watchObservedRunningTime="2026-01-06 14:38:20.029694254 +0000 UTC m=+96.657160602" Jan 06 14:38:21 crc kubenswrapper[4744]: I0106 14:38:21.027049 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4xw2" event={"ID":"1d41f99a-8352-472d-9011-e9c0259abf7b","Type":"ContainerStarted","Data":"f954106252cf6a0d55f9257f4bf8b1f9f15de8e705d2e48d15855de94a630548"} Jan 06 14:38:21 crc kubenswrapper[4744]: I0106 14:38:21.056590 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m4xw2" podStartSLOduration=3.485667217 podStartE2EDuration="1m0.056572289s" podCreationTimestamp="2026-01-06 14:37:21 +0000 UTC" firstStartedPulling="2026-01-06 14:37:23.865718706 +0000 UTC m=+40.493185024" lastFinishedPulling="2026-01-06 14:38:20.436623788 +0000 UTC m=+97.064090096" observedRunningTime="2026-01-06 14:38:21.052610839 +0000 UTC m=+97.680077157" watchObservedRunningTime="2026-01-06 14:38:21.056572289 +0000 UTC m=+97.684038627" Jan 06 14:38:21 crc kubenswrapper[4744]: I0106 14:38:21.095692 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:38:21 crc kubenswrapper[4744]: I0106 14:38:21.137027 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:38:21 crc kubenswrapper[4744]: I0106 14:38:21.428623 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:38:21 crc kubenswrapper[4744]: I0106 14:38:21.475799 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:38:21 crc kubenswrapper[4744]: I0106 14:38:21.563682 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:38:21 crc kubenswrapper[4744]: I0106 14:38:21.563735 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:38:22 crc kubenswrapper[4744]: I0106 14:38:22.600593 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-m4xw2" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerName="registry-server" probeResult="failure" output=< Jan 06 14:38:22 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:38:22 crc kubenswrapper[4744]: > Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.038781 4744 generic.go:334] "Generic (PLEG): container finished" podID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerID="56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176" exitCode=0 Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.038864 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hxcm" event={"ID":"6fe6e90c-e807-441c-b130-15a7368b87f0","Type":"ContainerDied","Data":"56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176"} Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.043651 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69twj" event={"ID":"6cbba207-bda4-4f95-a8f1-d003787cb231","Type":"ContainerStarted","Data":"21a686d85fd597d37d2832f046be2d0e925edbbd0b2950c79a21a6ebd7e9ed27"} Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.047028 4744 generic.go:334] "Generic (PLEG): container finished" podID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerID="3efda0bf5b56d5fc8f02d63ac2747b178793e26cd17792b2cf15bc2b6cae2024" exitCode=0 Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.047054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv9qm" event={"ID":"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe","Type":"ContainerDied","Data":"3efda0bf5b56d5fc8f02d63ac2747b178793e26cd17792b2cf15bc2b6cae2024"} Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.259056 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j999h"] Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.259744 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j999h" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerName="registry-server" containerID="cri-o://ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d" gracePeriod=2 Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.419699 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9z7kv"] Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.899388 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.950388 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tfgc\" (UniqueName: \"kubernetes.io/projected/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-kube-api-access-2tfgc\") pod \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " Jan 06 14:38:23 crc kubenswrapper[4744]: I0106 14:38:23.958669 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-kube-api-access-2tfgc" (OuterVolumeSpecName: "kube-api-access-2tfgc") pod "e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" (UID: "e7192120-5ddb-4ca8-a901-7ae9ebe22bdf"). InnerVolumeSpecName "kube-api-access-2tfgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.051941 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-catalog-content\") pod \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.051990 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-utilities\") pod \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\" (UID: \"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf\") " Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.052327 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tfgc\" (UniqueName: \"kubernetes.io/projected/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-kube-api-access-2tfgc\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.053139 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-utilities" (OuterVolumeSpecName: "utilities") pod "e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" (UID: "e7192120-5ddb-4ca8-a901-7ae9ebe22bdf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.054336 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bh4q" event={"ID":"408386a1-c1c3-4b95-8bce-4c60433159a0","Type":"ContainerStarted","Data":"66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5"} Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.055912 4744 generic.go:334] "Generic (PLEG): container finished" podID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerID="ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d" exitCode=0 Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.055976 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j999h" event={"ID":"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf","Type":"ContainerDied","Data":"ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d"} Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.055990 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j999h" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.056002 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j999h" event={"ID":"e7192120-5ddb-4ca8-a901-7ae9ebe22bdf","Type":"ContainerDied","Data":"f3d39d2f6b25f93b055c1c8de6492a2e44703a12949b3f82b331050d3fcae806"} Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.056018 4744 scope.go:117] "RemoveContainer" containerID="ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.057718 4744 generic.go:334] "Generic (PLEG): container finished" podID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerID="21a686d85fd597d37d2832f046be2d0e925edbbd0b2950c79a21a6ebd7e9ed27" exitCode=0 Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.057756 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69twj" event={"ID":"6cbba207-bda4-4f95-a8f1-d003787cb231","Type":"ContainerDied","Data":"21a686d85fd597d37d2832f046be2d0e925edbbd0b2950c79a21a6ebd7e9ed27"} Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.061350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hxcm" event={"ID":"6fe6e90c-e807-441c-b130-15a7368b87f0","Type":"ContainerStarted","Data":"e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380"} Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.067290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv9qm" event={"ID":"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe","Type":"ContainerStarted","Data":"dd331696e0622695dcffcbd582cf34fe29cec12e719fae276b5622bba7b6703d"} Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.076104 4744 scope.go:117] "RemoveContainer" containerID="ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.092671 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4hxcm" podStartSLOduration=3.798255345 podStartE2EDuration="1m2.092656582s" podCreationTimestamp="2026-01-06 14:37:22 +0000 UTC" firstStartedPulling="2026-01-06 14:37:25.227210985 +0000 UTC m=+41.854677303" lastFinishedPulling="2026-01-06 14:38:23.521612222 +0000 UTC m=+100.149078540" observedRunningTime="2026-01-06 14:38:24.091411197 +0000 UTC m=+100.718877515" watchObservedRunningTime="2026-01-06 14:38:24.092656582 +0000 UTC m=+100.720122900" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.095405 4744 scope.go:117] "RemoveContainer" containerID="491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.109220 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fv9qm" podStartSLOduration=2.534465653 podStartE2EDuration="1m1.109204792s" podCreationTimestamp="2026-01-06 14:37:23 +0000 UTC" firstStartedPulling="2026-01-06 14:37:25.228382366 +0000 UTC m=+41.855848684" lastFinishedPulling="2026-01-06 14:38:23.803121505 +0000 UTC m=+100.430587823" observedRunningTime="2026-01-06 14:38:24.106876497 +0000 UTC m=+100.734342815" watchObservedRunningTime="2026-01-06 14:38:24.109204792 +0000 UTC m=+100.736671110" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.115143 4744 scope.go:117] "RemoveContainer" containerID="ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d" Jan 06 14:38:24 crc kubenswrapper[4744]: E0106 14:38:24.115724 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d\": container with ID starting with ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d not found: ID does not exist" containerID="ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.115777 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d"} err="failed to get container status \"ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d\": rpc error: code = NotFound desc = could not find container \"ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d\": container with ID starting with ad7ec15a321f2825dd375ef52a4d8a256beddef14ee90dbac6b71b9ff281ea7d not found: ID does not exist" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.115801 4744 scope.go:117] "RemoveContainer" containerID="ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665" Jan 06 14:38:24 crc kubenswrapper[4744]: E0106 14:38:24.116274 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665\": container with ID starting with ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665 not found: ID does not exist" containerID="ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.116300 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665"} err="failed to get container status \"ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665\": rpc error: code = NotFound desc = could not find container \"ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665\": container with ID starting with ed388a225287ff1ff1d61bb23eb49204ca05285445a9a604c86c2acf904e6665 not found: ID does not exist" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.116318 4744 scope.go:117] "RemoveContainer" containerID="491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e" Jan 06 14:38:24 crc kubenswrapper[4744]: E0106 14:38:24.116593 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e\": container with ID starting with 491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e not found: ID does not exist" containerID="491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.116638 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e"} err="failed to get container status \"491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e\": rpc error: code = NotFound desc = could not find container \"491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e\": container with ID starting with 491b68eafde3ef9fa5ec5a40f35f0b9cdb0006a78404313ca96f2ba195a2e65e not found: ID does not exist" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.129079 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" (UID: "e7192120-5ddb-4ca8-a901-7ae9ebe22bdf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.153254 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.153430 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.440523 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j999h"] Jan 06 14:38:24 crc kubenswrapper[4744]: I0106 14:38:24.444295 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j999h"] Jan 06 14:38:25 crc kubenswrapper[4744]: I0106 14:38:25.075448 4744 generic.go:334] "Generic (PLEG): container finished" podID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerID="66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5" exitCode=0 Jan 06 14:38:25 crc kubenswrapper[4744]: I0106 14:38:25.075538 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bh4q" event={"ID":"408386a1-c1c3-4b95-8bce-4c60433159a0","Type":"ContainerDied","Data":"66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5"} Jan 06 14:38:25 crc kubenswrapper[4744]: I0106 14:38:25.717782 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" path="/var/lib/kubelet/pods/e7192120-5ddb-4ca8-a901-7ae9ebe22bdf/volumes" Jan 06 14:38:26 crc kubenswrapper[4744]: I0106 14:38:26.090597 4744 generic.go:334] "Generic (PLEG): container finished" podID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerID="2ecf8435a9157687bd7b0e63b5a0528dcef2691ab79a21eeca329addb84772c0" exitCode=0 Jan 06 14:38:26 crc kubenswrapper[4744]: I0106 14:38:26.090695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cwvb" event={"ID":"14d7cfd7-181f-4d79-8951-27d300d49a8f","Type":"ContainerDied","Data":"2ecf8435a9157687bd7b0e63b5a0528dcef2691ab79a21eeca329addb84772c0"} Jan 06 14:38:26 crc kubenswrapper[4744]: I0106 14:38:26.093891 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69twj" event={"ID":"6cbba207-bda4-4f95-a8f1-d003787cb231","Type":"ContainerStarted","Data":"03c57d078d71615f68e0f52deec8dd4822f04420426d672d0a8c913bce2be1b7"} Jan 06 14:38:26 crc kubenswrapper[4744]: I0106 14:38:26.130234 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-69twj" podStartSLOduration=4.038416822 podStartE2EDuration="1m3.130210389s" podCreationTimestamp="2026-01-06 14:37:23 +0000 UTC" firstStartedPulling="2026-01-06 14:37:26.210250937 +0000 UTC m=+42.837717255" lastFinishedPulling="2026-01-06 14:38:25.302044504 +0000 UTC m=+101.929510822" observedRunningTime="2026-01-06 14:38:26.128335697 +0000 UTC m=+102.755802015" watchObservedRunningTime="2026-01-06 14:38:26.130210389 +0000 UTC m=+102.757676717" Jan 06 14:38:27 crc kubenswrapper[4744]: I0106 14:38:27.101180 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bh4q" event={"ID":"408386a1-c1c3-4b95-8bce-4c60433159a0","Type":"ContainerStarted","Data":"64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1"} Jan 06 14:38:27 crc kubenswrapper[4744]: I0106 14:38:27.126138 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4bh4q" podStartSLOduration=4.887425666 podStartE2EDuration="1m4.126120522s" podCreationTimestamp="2026-01-06 14:37:23 +0000 UTC" firstStartedPulling="2026-01-06 14:37:26.204585298 +0000 UTC m=+42.832051616" lastFinishedPulling="2026-01-06 14:38:25.443280154 +0000 UTC m=+102.070746472" observedRunningTime="2026-01-06 14:38:27.122995205 +0000 UTC m=+103.750461543" watchObservedRunningTime="2026-01-06 14:38:27.126120522 +0000 UTC m=+103.753586860" Jan 06 14:38:28 crc kubenswrapper[4744]: I0106 14:38:28.108475 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cwvb" event={"ID":"14d7cfd7-181f-4d79-8951-27d300d49a8f","Type":"ContainerStarted","Data":"f79f3e3e7070ccaaf2c98dba729e886d3e24853359fa99139a27d0a23a1d41ac"} Jan 06 14:38:28 crc kubenswrapper[4744]: I0106 14:38:28.127314 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4cwvb" podStartSLOduration=4.262608394 podStartE2EDuration="1m8.127294541s" podCreationTimestamp="2026-01-06 14:37:20 +0000 UTC" firstStartedPulling="2026-01-06 14:37:23.5336741 +0000 UTC m=+40.161140418" lastFinishedPulling="2026-01-06 14:38:27.398360207 +0000 UTC m=+104.025826565" observedRunningTime="2026-01-06 14:38:28.123434074 +0000 UTC m=+104.750900402" watchObservedRunningTime="2026-01-06 14:38:28.127294541 +0000 UTC m=+104.754760869" Jan 06 14:38:31 crc kubenswrapper[4744]: I0106 14:38:31.179462 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:38:31 crc kubenswrapper[4744]: I0106 14:38:31.180244 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:38:31 crc kubenswrapper[4744]: I0106 14:38:31.229727 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:38:31 crc kubenswrapper[4744]: I0106 14:38:31.612501 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:38:31 crc kubenswrapper[4744]: I0106 14:38:31.670192 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:38:32 crc kubenswrapper[4744]: I0106 14:38:32.199837 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:38:32 crc kubenswrapper[4744]: I0106 14:38:32.463009 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m4xw2"] Jan 06 14:38:33 crc kubenswrapper[4744]: I0106 14:38:33.138496 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m4xw2" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerName="registry-server" containerID="cri-o://f954106252cf6a0d55f9257f4bf8b1f9f15de8e705d2e48d15855de94a630548" gracePeriod=2 Jan 06 14:38:33 crc kubenswrapper[4744]: I0106 14:38:33.164791 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:38:33 crc kubenswrapper[4744]: I0106 14:38:33.165502 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:38:33 crc kubenswrapper[4744]: I0106 14:38:33.225039 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:38:33 crc kubenswrapper[4744]: I0106 14:38:33.600993 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:38:33 crc kubenswrapper[4744]: I0106 14:38:33.601069 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:38:33 crc kubenswrapper[4744]: I0106 14:38:33.661265 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:38:34 crc kubenswrapper[4744]: I0106 14:38:34.199512 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:38:34 crc kubenswrapper[4744]: I0106 14:38:34.205323 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:38:34 crc kubenswrapper[4744]: I0106 14:38:34.698453 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:38:34 crc kubenswrapper[4744]: I0106 14:38:34.698497 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:38:34 crc kubenswrapper[4744]: I0106 14:38:34.698509 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:38:34 crc kubenswrapper[4744]: I0106 14:38:34.698517 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:38:34 crc kubenswrapper[4744]: I0106 14:38:34.745413 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.151365 4744 generic.go:334] "Generic (PLEG): container finished" podID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerID="f954106252cf6a0d55f9257f4bf8b1f9f15de8e705d2e48d15855de94a630548" exitCode=0 Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.151464 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4xw2" event={"ID":"1d41f99a-8352-472d-9011-e9c0259abf7b","Type":"ContainerDied","Data":"f954106252cf6a0d55f9257f4bf8b1f9f15de8e705d2e48d15855de94a630548"} Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.187440 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.459475 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv9qm"] Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.744987 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4bh4q" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerName="registry-server" probeResult="failure" output=< Jan 06 14:38:35 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:38:35 crc kubenswrapper[4744]: > Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.777235 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.921312 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-utilities\") pod \"1d41f99a-8352-472d-9011-e9c0259abf7b\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.921871 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-catalog-content\") pod \"1d41f99a-8352-472d-9011-e9c0259abf7b\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.921984 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49v7g\" (UniqueName: \"kubernetes.io/projected/1d41f99a-8352-472d-9011-e9c0259abf7b-kube-api-access-49v7g\") pod \"1d41f99a-8352-472d-9011-e9c0259abf7b\" (UID: \"1d41f99a-8352-472d-9011-e9c0259abf7b\") " Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.923286 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-utilities" (OuterVolumeSpecName: "utilities") pod "1d41f99a-8352-472d-9011-e9c0259abf7b" (UID: "1d41f99a-8352-472d-9011-e9c0259abf7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.935487 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d41f99a-8352-472d-9011-e9c0259abf7b-kube-api-access-49v7g" (OuterVolumeSpecName: "kube-api-access-49v7g") pod "1d41f99a-8352-472d-9011-e9c0259abf7b" (UID: "1d41f99a-8352-472d-9011-e9c0259abf7b"). InnerVolumeSpecName "kube-api-access-49v7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:35 crc kubenswrapper[4744]: I0106 14:38:35.996781 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d41f99a-8352-472d-9011-e9c0259abf7b" (UID: "1d41f99a-8352-472d-9011-e9c0259abf7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.023375 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.023615 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d41f99a-8352-472d-9011-e9c0259abf7b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.023679 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49v7g\" (UniqueName: \"kubernetes.io/projected/1d41f99a-8352-472d-9011-e9c0259abf7b-kube-api-access-49v7g\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.163032 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m4xw2" event={"ID":"1d41f99a-8352-472d-9011-e9c0259abf7b","Type":"ContainerDied","Data":"59f4c5028f5391f4cc262c987ca6c785281778a81d20abdca3572e3dd00d0cf6"} Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.163384 4744 scope.go:117] "RemoveContainer" containerID="f954106252cf6a0d55f9257f4bf8b1f9f15de8e705d2e48d15855de94a630548" Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.163644 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m4xw2" Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.164789 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fv9qm" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerName="registry-server" containerID="cri-o://dd331696e0622695dcffcbd582cf34fe29cec12e719fae276b5622bba7b6703d" gracePeriod=2 Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.195986 4744 scope.go:117] "RemoveContainer" containerID="c1b039cd6c94530f4507938fcb9cd6902a45038c22828cd02ec0df9dc77c190f" Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.210598 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m4xw2"] Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.213455 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m4xw2"] Jan 06 14:38:36 crc kubenswrapper[4744]: I0106 14:38:36.226345 4744 scope.go:117] "RemoveContainer" containerID="395f98bd0eac1daa06856550ce62970d259464d9e5bd0c8308807d150879896e" Jan 06 14:38:37 crc kubenswrapper[4744]: I0106 14:38:37.719843 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" path="/var/lib/kubelet/pods/1d41f99a-8352-472d-9011-e9c0259abf7b/volumes" Jan 06 14:38:38 crc kubenswrapper[4744]: I0106 14:38:38.192415 4744 generic.go:334] "Generic (PLEG): container finished" podID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerID="dd331696e0622695dcffcbd582cf34fe29cec12e719fae276b5622bba7b6703d" exitCode=0 Jan 06 14:38:38 crc kubenswrapper[4744]: I0106 14:38:38.192480 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv9qm" event={"ID":"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe","Type":"ContainerDied","Data":"dd331696e0622695dcffcbd582cf34fe29cec12e719fae276b5622bba7b6703d"} Jan 06 14:38:38 crc kubenswrapper[4744]: I0106 14:38:38.465917 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-69twj"] Jan 06 14:38:38 crc kubenswrapper[4744]: I0106 14:38:38.466985 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-69twj" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerName="registry-server" containerID="cri-o://03c57d078d71615f68e0f52deec8dd4822f04420426d672d0a8c913bce2be1b7" gracePeriod=2 Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.332283 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.473567 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcvmd\" (UniqueName: \"kubernetes.io/projected/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-kube-api-access-wcvmd\") pod \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.473732 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-utilities\") pod \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.473776 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-catalog-content\") pod \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\" (UID: \"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe\") " Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.475143 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-utilities" (OuterVolumeSpecName: "utilities") pod "b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" (UID: "b1d9691a-454d-4ec0-9447-5cf0df1f5cfe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.478351 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-kube-api-access-wcvmd" (OuterVolumeSpecName: "kube-api-access-wcvmd") pod "b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" (UID: "b1d9691a-454d-4ec0-9447-5cf0df1f5cfe"). InnerVolumeSpecName "kube-api-access-wcvmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.505345 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" (UID: "b1d9691a-454d-4ec0-9447-5cf0df1f5cfe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.575505 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.575545 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:39 crc kubenswrapper[4744]: I0106 14:38:39.575559 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcvmd\" (UniqueName: \"kubernetes.io/projected/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe-kube-api-access-wcvmd\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:40 crc kubenswrapper[4744]: I0106 14:38:40.209930 4744 generic.go:334] "Generic (PLEG): container finished" podID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerID="03c57d078d71615f68e0f52deec8dd4822f04420426d672d0a8c913bce2be1b7" exitCode=0 Jan 06 14:38:40 crc kubenswrapper[4744]: I0106 14:38:40.210025 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69twj" event={"ID":"6cbba207-bda4-4f95-a8f1-d003787cb231","Type":"ContainerDied","Data":"03c57d078d71615f68e0f52deec8dd4822f04420426d672d0a8c913bce2be1b7"} Jan 06 14:38:40 crc kubenswrapper[4744]: I0106 14:38:40.212995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fv9qm" event={"ID":"b1d9691a-454d-4ec0-9447-5cf0df1f5cfe","Type":"ContainerDied","Data":"dbe15818a0e63f9e6998d0dca78bd5410a481811f0f4db4c5c2697858393c7b9"} Jan 06 14:38:40 crc kubenswrapper[4744]: I0106 14:38:40.213039 4744 scope.go:117] "RemoveContainer" containerID="dd331696e0622695dcffcbd582cf34fe29cec12e719fae276b5622bba7b6703d" Jan 06 14:38:40 crc kubenswrapper[4744]: I0106 14:38:40.213127 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fv9qm" Jan 06 14:38:40 crc kubenswrapper[4744]: I0106 14:38:40.237112 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv9qm"] Jan 06 14:38:40 crc kubenswrapper[4744]: I0106 14:38:40.239244 4744 scope.go:117] "RemoveContainer" containerID="3efda0bf5b56d5fc8f02d63ac2747b178793e26cd17792b2cf15bc2b6cae2024" Jan 06 14:38:40 crc kubenswrapper[4744]: I0106 14:38:40.243232 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fv9qm"] Jan 06 14:38:40 crc kubenswrapper[4744]: I0106 14:38:40.265400 4744 scope.go:117] "RemoveContainer" containerID="5625d7cac6383b53c66ad4cd0c6a3470229ca31d4696074fa954b0c65522dde6" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.148666 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.220021 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-69twj" event={"ID":"6cbba207-bda4-4f95-a8f1-d003787cb231","Type":"ContainerDied","Data":"f967f280b5a6dad613e78688c20fa20a567e138e768b7cb4241437bebe32aab2"} Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.220077 4744 scope.go:117] "RemoveContainer" containerID="03c57d078d71615f68e0f52deec8dd4822f04420426d672d0a8c913bce2be1b7" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.220232 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-69twj" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.235622 4744 scope.go:117] "RemoveContainer" containerID="21a686d85fd597d37d2832f046be2d0e925edbbd0b2950c79a21a6ebd7e9ed27" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.249312 4744 scope.go:117] "RemoveContainer" containerID="15b52fb24e6d33d4e904e1c5359b148cf27973de6b1ea034f7e0004c2b646ca0" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.303406 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-catalog-content\") pod \"6cbba207-bda4-4f95-a8f1-d003787cb231\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.303466 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbt7\" (UniqueName: \"kubernetes.io/projected/6cbba207-bda4-4f95-a8f1-d003787cb231-kube-api-access-jhbt7\") pod \"6cbba207-bda4-4f95-a8f1-d003787cb231\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.303495 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-utilities\") pod \"6cbba207-bda4-4f95-a8f1-d003787cb231\" (UID: \"6cbba207-bda4-4f95-a8f1-d003787cb231\") " Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.304916 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-utilities" (OuterVolumeSpecName: "utilities") pod "6cbba207-bda4-4f95-a8f1-d003787cb231" (UID: "6cbba207-bda4-4f95-a8f1-d003787cb231"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.309375 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cbba207-bda4-4f95-a8f1-d003787cb231-kube-api-access-jhbt7" (OuterVolumeSpecName: "kube-api-access-jhbt7") pod "6cbba207-bda4-4f95-a8f1-d003787cb231" (UID: "6cbba207-bda4-4f95-a8f1-d003787cb231"). InnerVolumeSpecName "kube-api-access-jhbt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.404820 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbt7\" (UniqueName: \"kubernetes.io/projected/6cbba207-bda4-4f95-a8f1-d003787cb231-kube-api-access-jhbt7\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.404854 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.473777 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6cbba207-bda4-4f95-a8f1-d003787cb231" (UID: "6cbba207-bda4-4f95-a8f1-d003787cb231"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.506226 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cbba207-bda4-4f95-a8f1-d003787cb231-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.553953 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-69twj"] Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.557559 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-69twj"] Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.718376 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" path="/var/lib/kubelet/pods/6cbba207-bda4-4f95-a8f1-d003787cb231/volumes" Jan 06 14:38:41 crc kubenswrapper[4744]: I0106 14:38:41.718959 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" path="/var/lib/kubelet/pods/b1d9691a-454d-4ec0-9447-5cf0df1f5cfe/volumes" Jan 06 14:38:43 crc kubenswrapper[4744]: I0106 14:38:43.216989 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-86f674b85d-l6b67"] Jan 06 14:38:43 crc kubenswrapper[4744]: I0106 14:38:43.217464 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" podUID="5f926310-2075-4937-b88e-b6fd6184ef70" containerName="controller-manager" containerID="cri-o://7438288f53f0a5ac41617027d4316717a98e0e64d9053d58749cf5b2b9bdf0b2" gracePeriod=30 Jan 06 14:38:43 crc kubenswrapper[4744]: I0106 14:38:43.296667 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5"] Jan 06 14:38:43 crc kubenswrapper[4744]: I0106 14:38:43.296892 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" podUID="7273d766-2526-4fbe-836a-3af7c3eb37cf" containerName="route-controller-manager" containerID="cri-o://bfd235c30e5fb7cf3a5af31593e5d836a8220dd40b1c88a86e297e8246adb6c2" gracePeriod=30 Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.247047 4744 generic.go:334] "Generic (PLEG): container finished" podID="5f926310-2075-4937-b88e-b6fd6184ef70" containerID="7438288f53f0a5ac41617027d4316717a98e0e64d9053d58749cf5b2b9bdf0b2" exitCode=0 Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.247109 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" event={"ID":"5f926310-2075-4937-b88e-b6fd6184ef70","Type":"ContainerDied","Data":"7438288f53f0a5ac41617027d4316717a98e0e64d9053d58749cf5b2b9bdf0b2"} Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.249901 4744 generic.go:334] "Generic (PLEG): container finished" podID="7273d766-2526-4fbe-836a-3af7c3eb37cf" containerID="bfd235c30e5fb7cf3a5af31593e5d836a8220dd40b1c88a86e297e8246adb6c2" exitCode=0 Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.249955 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" event={"ID":"7273d766-2526-4fbe-836a-3af7c3eb37cf","Type":"ContainerDied","Data":"bfd235c30e5fb7cf3a5af31593e5d836a8220dd40b1c88a86e297e8246adb6c2"} Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.765686 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.830432 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.942809 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969363 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl"] Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969547 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969558 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969567 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerName="extract-utilities" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969573 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerName="extract-utilities" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969584 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969591 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969599 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerName="extract-utilities" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969605 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerName="extract-utilities" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969616 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerName="extract-content" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969622 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerName="extract-content" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969632 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerName="extract-content" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969638 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerName="extract-content" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969646 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7273d766-2526-4fbe-836a-3af7c3eb37cf" containerName="route-controller-manager" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969651 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7273d766-2526-4fbe-836a-3af7c3eb37cf" containerName="route-controller-manager" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969659 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969665 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969671 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerName="extract-utilities" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969677 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerName="extract-utilities" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969684 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969690 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969698 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerName="extract-content" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969714 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerName="extract-content" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969724 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerName="extract-utilities" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969730 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerName="extract-utilities" Jan 06 14:38:44 crc kubenswrapper[4744]: E0106 14:38:44.969740 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerName="extract-content" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969746 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerName="extract-content" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969822 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d9691a-454d-4ec0-9447-5cf0df1f5cfe" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969833 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cbba207-bda4-4f95-a8f1-d003787cb231" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969840 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7273d766-2526-4fbe-836a-3af7c3eb37cf" containerName="route-controller-manager" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969847 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d41f99a-8352-472d-9011-e9c0259abf7b" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.969856 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7192120-5ddb-4ca8-a901-7ae9ebe22bdf" containerName="registry-server" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.970235 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.985027 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:44 crc kubenswrapper[4744]: I0106 14:38:44.996415 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl"] Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.058383 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-config\") pod \"7273d766-2526-4fbe-836a-3af7c3eb37cf\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.058432 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-client-ca\") pod \"7273d766-2526-4fbe-836a-3af7c3eb37cf\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.058547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldk9l\" (UniqueName: \"kubernetes.io/projected/7273d766-2526-4fbe-836a-3af7c3eb37cf-kube-api-access-ldk9l\") pod \"7273d766-2526-4fbe-836a-3af7c3eb37cf\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.058607 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7273d766-2526-4fbe-836a-3af7c3eb37cf-serving-cert\") pod \"7273d766-2526-4fbe-836a-3af7c3eb37cf\" (UID: \"7273d766-2526-4fbe-836a-3af7c3eb37cf\") " Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.058783 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-client-ca\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.058831 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7046fa2e-834b-4802-a76c-d70de3855872-serving-cert\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.058906 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75fj5\" (UniqueName: \"kubernetes.io/projected/7046fa2e-834b-4802-a76c-d70de3855872-kube-api-access-75fj5\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.058970 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-config\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.059263 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-client-ca" (OuterVolumeSpecName: "client-ca") pod "7273d766-2526-4fbe-836a-3af7c3eb37cf" (UID: "7273d766-2526-4fbe-836a-3af7c3eb37cf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.059371 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-config" (OuterVolumeSpecName: "config") pod "7273d766-2526-4fbe-836a-3af7c3eb37cf" (UID: "7273d766-2526-4fbe-836a-3af7c3eb37cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.067275 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7273d766-2526-4fbe-836a-3af7c3eb37cf-kube-api-access-ldk9l" (OuterVolumeSpecName: "kube-api-access-ldk9l") pod "7273d766-2526-4fbe-836a-3af7c3eb37cf" (UID: "7273d766-2526-4fbe-836a-3af7c3eb37cf"). InnerVolumeSpecName "kube-api-access-ldk9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.069269 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7273d766-2526-4fbe-836a-3af7c3eb37cf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7273d766-2526-4fbe-836a-3af7c3eb37cf" (UID: "7273d766-2526-4fbe-836a-3af7c3eb37cf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f926310-2075-4937-b88e-b6fd6184ef70-serving-cert\") pod \"5f926310-2075-4937-b88e-b6fd6184ef70\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159456 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-config\") pod \"5f926310-2075-4937-b88e-b6fd6184ef70\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159487 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-client-ca\") pod \"5f926310-2075-4937-b88e-b6fd6184ef70\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159518 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdd9t\" (UniqueName: \"kubernetes.io/projected/5f926310-2075-4937-b88e-b6fd6184ef70-kube-api-access-wdd9t\") pod \"5f926310-2075-4937-b88e-b6fd6184ef70\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159555 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-proxy-ca-bundles\") pod \"5f926310-2075-4937-b88e-b6fd6184ef70\" (UID: \"5f926310-2075-4937-b88e-b6fd6184ef70\") " Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159657 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-client-ca\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7046fa2e-834b-4802-a76c-d70de3855872-serving-cert\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159719 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75fj5\" (UniqueName: \"kubernetes.io/projected/7046fa2e-834b-4802-a76c-d70de3855872-kube-api-access-75fj5\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159759 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-config\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159801 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159811 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7273d766-2526-4fbe-836a-3af7c3eb37cf-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159820 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldk9l\" (UniqueName: \"kubernetes.io/projected/7273d766-2526-4fbe-836a-3af7c3eb37cf-kube-api-access-ldk9l\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.159830 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7273d766-2526-4fbe-836a-3af7c3eb37cf-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.160813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-config\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.161478 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5f926310-2075-4937-b88e-b6fd6184ef70" (UID: "5f926310-2075-4937-b88e-b6fd6184ef70"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.161933 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-config" (OuterVolumeSpecName: "config") pod "5f926310-2075-4937-b88e-b6fd6184ef70" (UID: "5f926310-2075-4937-b88e-b6fd6184ef70"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.162355 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-client-ca" (OuterVolumeSpecName: "client-ca") pod "5f926310-2075-4937-b88e-b6fd6184ef70" (UID: "5f926310-2075-4937-b88e-b6fd6184ef70"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.163883 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-client-ca\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.165061 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7046fa2e-834b-4802-a76c-d70de3855872-serving-cert\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.166354 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f926310-2075-4937-b88e-b6fd6184ef70-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5f926310-2075-4937-b88e-b6fd6184ef70" (UID: "5f926310-2075-4937-b88e-b6fd6184ef70"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.166356 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f926310-2075-4937-b88e-b6fd6184ef70-kube-api-access-wdd9t" (OuterVolumeSpecName: "kube-api-access-wdd9t") pod "5f926310-2075-4937-b88e-b6fd6184ef70" (UID: "5f926310-2075-4937-b88e-b6fd6184ef70"). InnerVolumeSpecName "kube-api-access-wdd9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.184687 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75fj5\" (UniqueName: \"kubernetes.io/projected/7046fa2e-834b-4802-a76c-d70de3855872-kube-api-access-75fj5\") pod \"route-controller-manager-6f5bb76d5c-slktl\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.256849 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.259320 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f674b85d-l6b67" event={"ID":"5f926310-2075-4937-b88e-b6fd6184ef70","Type":"ContainerDied","Data":"b9bdccf2cbf585eacebdf034a8959afee01236284eaaca4482183a808127d58c"} Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.259390 4744 scope.go:117] "RemoveContainer" containerID="7438288f53f0a5ac41617027d4316717a98e0e64d9053d58749cf5b2b9bdf0b2" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.260654 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.260681 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.260692 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdd9t\" (UniqueName: \"kubernetes.io/projected/5f926310-2075-4937-b88e-b6fd6184ef70-kube-api-access-wdd9t\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.260704 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f926310-2075-4937-b88e-b6fd6184ef70-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.260712 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f926310-2075-4937-b88e-b6fd6184ef70-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.262323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" event={"ID":"7273d766-2526-4fbe-836a-3af7c3eb37cf","Type":"ContainerDied","Data":"8920aba6571811643b97da00779d64b7212b684cd06d16eb9dca3594e2bc263f"} Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.263387 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.277100 4744 scope.go:117] "RemoveContainer" containerID="bfd235c30e5fb7cf3a5af31593e5d836a8220dd40b1c88a86e297e8246adb6c2" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.289312 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-86f674b85d-l6b67"] Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.292630 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-86f674b85d-l6b67"] Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.298256 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.307663 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5"] Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.318444 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796798f6b5-jm7x5"] Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.531609 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl"] Jan 06 14:38:45 crc kubenswrapper[4744]: W0106 14:38:45.545501 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7046fa2e_834b_4802_a76c_d70de3855872.slice/crio-8c84fd0011397647c33b7b32c28d6b89e808407fb93abfe6ed92ef8d5a24ecd3 WatchSource:0}: Error finding container 8c84fd0011397647c33b7b32c28d6b89e808407fb93abfe6ed92ef8d5a24ecd3: Status 404 returned error can't find the container with id 8c84fd0011397647c33b7b32c28d6b89e808407fb93abfe6ed92ef8d5a24ecd3 Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.719379 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f926310-2075-4937-b88e-b6fd6184ef70" path="/var/lib/kubelet/pods/5f926310-2075-4937-b88e-b6fd6184ef70/volumes" Jan 06 14:38:45 crc kubenswrapper[4744]: I0106 14:38:45.720780 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7273d766-2526-4fbe-836a-3af7c3eb37cf" path="/var/lib/kubelet/pods/7273d766-2526-4fbe-836a-3af7c3eb37cf/volumes" Jan 06 14:38:46 crc kubenswrapper[4744]: I0106 14:38:46.268427 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" event={"ID":"7046fa2e-834b-4802-a76c-d70de3855872","Type":"ContainerStarted","Data":"8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a"} Jan 06 14:38:46 crc kubenswrapper[4744]: I0106 14:38:46.268479 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" event={"ID":"7046fa2e-834b-4802-a76c-d70de3855872","Type":"ContainerStarted","Data":"8c84fd0011397647c33b7b32c28d6b89e808407fb93abfe6ed92ef8d5a24ecd3"} Jan 06 14:38:46 crc kubenswrapper[4744]: I0106 14:38:46.268502 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:46 crc kubenswrapper[4744]: I0106 14:38:46.289556 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" podStartSLOduration=3.289540053 podStartE2EDuration="3.289540053s" podCreationTimestamp="2026-01-06 14:38:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:38:46.285709897 +0000 UTC m=+122.913176225" watchObservedRunningTime="2026-01-06 14:38:46.289540053 +0000 UTC m=+122.917006391" Jan 06 14:38:46 crc kubenswrapper[4744]: I0106 14:38:46.554705 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.921144 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b68d4bc88-v86n5"] Jan 06 14:38:47 crc kubenswrapper[4744]: E0106 14:38:47.921777 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f926310-2075-4937-b88e-b6fd6184ef70" containerName="controller-manager" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.921798 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f926310-2075-4937-b88e-b6fd6184ef70" containerName="controller-manager" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.921971 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f926310-2075-4937-b88e-b6fd6184ef70" containerName="controller-manager" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.922559 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.929240 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.929335 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.930470 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.932876 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.933429 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.933928 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.949617 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b68d4bc88-v86n5"] Jan 06 14:38:47 crc kubenswrapper[4744]: I0106 14:38:47.962530 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.102753 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-proxy-ca-bundles\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.102848 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfg5c\" (UniqueName: \"kubernetes.io/projected/360f121c-7279-4311-9331-3abf4d0dd528-kube-api-access-hfg5c\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.102997 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-client-ca\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.103097 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/360f121c-7279-4311-9331-3abf4d0dd528-serving-cert\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.103191 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-config\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.205067 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-proxy-ca-bundles\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.205224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfg5c\" (UniqueName: \"kubernetes.io/projected/360f121c-7279-4311-9331-3abf4d0dd528-kube-api-access-hfg5c\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.205320 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-client-ca\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.205418 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/360f121c-7279-4311-9331-3abf4d0dd528-serving-cert\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.205510 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-config\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.207247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-client-ca\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.208115 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-config\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.209030 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-proxy-ca-bundles\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.221739 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/360f121c-7279-4311-9331-3abf4d0dd528-serving-cert\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.245375 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfg5c\" (UniqueName: \"kubernetes.io/projected/360f121c-7279-4311-9331-3abf4d0dd528-kube-api-access-hfg5c\") pod \"controller-manager-6b68d4bc88-v86n5\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.253754 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.462786 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" podUID="82c5f022-eee7-42a1-b9b3-e00c07783034" containerName="oauth-openshift" containerID="cri-o://93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263" gracePeriod=15 Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.761752 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b68d4bc88-v86n5"] Jan 06 14:38:48 crc kubenswrapper[4744]: W0106 14:38:48.772805 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod360f121c_7279_4311_9331_3abf4d0dd528.slice/crio-e1536443832322e9c236dcdf1a6ca0fed98181ffe223b5245b786331971ab139 WatchSource:0}: Error finding container e1536443832322e9c236dcdf1a6ca0fed98181ffe223b5245b786331971ab139: Status 404 returned error can't find the container with id e1536443832322e9c236dcdf1a6ca0fed98181ffe223b5245b786331971ab139 Jan 06 14:38:48 crc kubenswrapper[4744]: I0106 14:38:48.863507 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017312 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-session\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017352 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-cliconfig\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017376 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017406 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017434 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017460 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017485 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017514 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gznpj\" (UniqueName: \"kubernetes.io/projected/82c5f022-eee7-42a1-b9b3-e00c07783034-kube-api-access-gznpj\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017540 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-router-certs\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017572 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-idp-0-file-data\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017603 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-dir\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017641 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.017667 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-error\") pod \"82c5f022-eee7-42a1-b9b3-e00c07783034\" (UID: \"82c5f022-eee7-42a1-b9b3-e00c07783034\") " Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.018731 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.018959 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.023835 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.024644 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.024878 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.025139 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.025948 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.026264 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.027423 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.027740 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.028139 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.028833 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.029281 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.029448 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c5f022-eee7-42a1-b9b3-e00c07783034-kube-api-access-gznpj" (OuterVolumeSpecName: "kube-api-access-gznpj") pod "82c5f022-eee7-42a1-b9b3-e00c07783034" (UID: "82c5f022-eee7-42a1-b9b3-e00c07783034"). InnerVolumeSpecName "kube-api-access-gznpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128621 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128666 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128682 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128694 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128711 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128722 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128736 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128748 4744 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128760 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gznpj\" (UniqueName: \"kubernetes.io/projected/82c5f022-eee7-42a1-b9b3-e00c07783034-kube-api-access-gznpj\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128773 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128786 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128797 4744 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/82c5f022-eee7-42a1-b9b3-e00c07783034-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128808 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.128820 4744 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/82c5f022-eee7-42a1-b9b3-e00c07783034-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.300867 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" event={"ID":"360f121c-7279-4311-9331-3abf4d0dd528","Type":"ContainerStarted","Data":"ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888"} Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.301190 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" event={"ID":"360f121c-7279-4311-9331-3abf4d0dd528","Type":"ContainerStarted","Data":"e1536443832322e9c236dcdf1a6ca0fed98181ffe223b5245b786331971ab139"} Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.301213 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.302293 4744 generic.go:334] "Generic (PLEG): container finished" podID="82c5f022-eee7-42a1-b9b3-e00c07783034" containerID="93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263" exitCode=0 Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.302331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" event={"ID":"82c5f022-eee7-42a1-b9b3-e00c07783034","Type":"ContainerDied","Data":"93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263"} Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.302365 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.302392 4744 scope.go:117] "RemoveContainer" containerID="93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.302380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-9z7kv" event={"ID":"82c5f022-eee7-42a1-b9b3-e00c07783034","Type":"ContainerDied","Data":"2755a17385d2b7d13e8fcb5b0e112663bf80b345942916b6bb75d37131b42f94"} Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.319816 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.321049 4744 scope.go:117] "RemoveContainer" containerID="93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263" Jan 06 14:38:49 crc kubenswrapper[4744]: E0106 14:38:49.321714 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263\": container with ID starting with 93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263 not found: ID does not exist" containerID="93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.321750 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263"} err="failed to get container status \"93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263\": rpc error: code = NotFound desc = could not find container \"93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263\": container with ID starting with 93f4610847a7e64c77f7193883bd02f51641e00bb4a74705871ba6839d4ea263 not found: ID does not exist" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.330345 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" podStartSLOduration=6.330321497 podStartE2EDuration="6.330321497s" podCreationTimestamp="2026-01-06 14:38:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:38:49.324201977 +0000 UTC m=+125.951668295" watchObservedRunningTime="2026-01-06 14:38:49.330321497 +0000 UTC m=+125.957787855" Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.364492 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9z7kv"] Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.364581 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-9z7kv"] Jan 06 14:38:49 crc kubenswrapper[4744]: I0106 14:38:49.719114 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82c5f022-eee7-42a1-b9b3-e00c07783034" path="/var/lib/kubelet/pods/82c5f022-eee7-42a1-b9b3-e00c07783034/volumes" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.769519 4744 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 06 14:38:56 crc kubenswrapper[4744]: E0106 14:38:56.770572 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c5f022-eee7-42a1-b9b3-e00c07783034" containerName="oauth-openshift" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.770600 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c5f022-eee7-42a1-b9b3-e00c07783034" containerName="oauth-openshift" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.770790 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c5f022-eee7-42a1-b9b3-e00c07783034" containerName="oauth-openshift" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.771494 4744 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.771906 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a" gracePeriod=15 Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.772108 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.772665 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5" gracePeriod=15 Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.772771 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776" gracePeriod=15 Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.772838 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885" gracePeriod=15 Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.772936 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1" gracePeriod=15 Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.773659 4744 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 06 14:38:56 crc kubenswrapper[4744]: E0106 14:38:56.773960 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.773987 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 06 14:38:56 crc kubenswrapper[4744]: E0106 14:38:56.774010 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774027 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 06 14:38:56 crc kubenswrapper[4744]: E0106 14:38:56.774045 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774060 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 06 14:38:56 crc kubenswrapper[4744]: E0106 14:38:56.774085 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774100 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 06 14:38:56 crc kubenswrapper[4744]: E0106 14:38:56.774124 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774140 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 06 14:38:56 crc kubenswrapper[4744]: E0106 14:38:56.774223 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774242 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 06 14:38:56 crc kubenswrapper[4744]: E0106 14:38:56.774267 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774285 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774515 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774545 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774572 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774589 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.774609 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.775009 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.816780 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.930487 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.930542 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.930571 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.930608 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.930676 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.930711 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.930738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:56 crc kubenswrapper[4744]: I0106 14:38:56.930763 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031725 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031795 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031828 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031856 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031868 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031891 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031936 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031949 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031878 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031967 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.031994 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.032056 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.032069 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.032085 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.114793 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:38:57 crc kubenswrapper[4744]: W0106 14:38:57.138342 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-b30ab850947817a7471b7a094b62dbd449b8625f9387fc2389921d59d5cd721e WatchSource:0}: Error finding container b30ab850947817a7471b7a094b62dbd449b8625f9387fc2389921d59d5cd721e: Status 404 returned error can't find the container with id b30ab850947817a7471b7a094b62dbd449b8625f9387fc2389921d59d5cd721e Jan 06 14:38:57 crc kubenswrapper[4744]: E0106 14:38:57.141845 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.22:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18882b2f970da538 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-06 14:38:57.14063084 +0000 UTC m=+133.768097178,LastTimestamp:2026-01-06 14:38:57.14063084 +0000 UTC m=+133.768097178,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.362517 4744 generic.go:334] "Generic (PLEG): container finished" podID="c7684584-6c37-4d30-8593-70d738b41d93" containerID="f3841de24ff12e431a5c37adf0bbe9b220965fc44b868a99ef4964b9c58de9cf" exitCode=0 Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.362580 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c7684584-6c37-4d30-8593-70d738b41d93","Type":"ContainerDied","Data":"f3841de24ff12e431a5c37adf0bbe9b220965fc44b868a99ef4964b9c58de9cf"} Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.363685 4744 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.363996 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.364318 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.369535 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.371243 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.372227 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5" exitCode=0 Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.372262 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776" exitCode=0 Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.372275 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885" exitCode=0 Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.372287 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1" exitCode=2 Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.372370 4744 scope.go:117] "RemoveContainer" containerID="30bc0343d6048ccb4cb2e7deb0c28359cfc62c1e0aa87ad993ad724d19bbad1b" Jan 06 14:38:57 crc kubenswrapper[4744]: I0106 14:38:57.373733 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b30ab850947817a7471b7a094b62dbd449b8625f9387fc2389921d59d5cd721e"} Jan 06 14:38:57 crc kubenswrapper[4744]: E0106 14:38:57.388917 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.22:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18882b2f970da538 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-06 14:38:57.14063084 +0000 UTC m=+133.768097178,LastTimestamp:2026-01-06 14:38:57.14063084 +0000 UTC m=+133.768097178,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.383058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f"} Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.383568 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.383957 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.385793 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.679558 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.680915 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.682025 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.879449 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-kubelet-dir\") pod \"c7684584-6c37-4d30-8593-70d738b41d93\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.879830 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-var-lock\") pod \"c7684584-6c37-4d30-8593-70d738b41d93\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.879958 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7684584-6c37-4d30-8593-70d738b41d93-kube-api-access\") pod \"c7684584-6c37-4d30-8593-70d738b41d93\" (UID: \"c7684584-6c37-4d30-8593-70d738b41d93\") " Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.881374 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c7684584-6c37-4d30-8593-70d738b41d93" (UID: "c7684584-6c37-4d30-8593-70d738b41d93"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.894711 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-var-lock" (OuterVolumeSpecName: "var-lock") pod "c7684584-6c37-4d30-8593-70d738b41d93" (UID: "c7684584-6c37-4d30-8593-70d738b41d93"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.913885 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7684584-6c37-4d30-8593-70d738b41d93-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c7684584-6c37-4d30-8593-70d738b41d93" (UID: "c7684584-6c37-4d30-8593-70d738b41d93"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.981701 4744 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.982080 4744 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c7684584-6c37-4d30-8593-70d738b41d93-var-lock\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:58 crc kubenswrapper[4744]: I0106 14:38:58.982100 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7684584-6c37-4d30-8593-70d738b41d93-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.129786 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.130657 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.131495 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.131822 4744 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.132083 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.184153 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.184492 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.184618 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.184287 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.184556 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.184644 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.185002 4744 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.185063 4744 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.185117 4744 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.397131 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.397584 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c7684584-6c37-4d30-8593-70d738b41d93","Type":"ContainerDied","Data":"04b6ad349bf54a1fd227cccd861666308d911b9fe43864e803c9ee5e831dbf7d"} Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.398601 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04b6ad349bf54a1fd227cccd861666308d911b9fe43864e803c9ee5e831dbf7d" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.401906 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.403242 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a" exitCode=0 Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.403323 4744 scope.go:117] "RemoveContainer" containerID="0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.403347 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.427019 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.427643 4744 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.427993 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.428537 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.428961 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.429300 4744 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.431130 4744 scope.go:117] "RemoveContainer" containerID="80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.453877 4744 scope.go:117] "RemoveContainer" containerID="6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.474706 4744 scope.go:117] "RemoveContainer" containerID="69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.492128 4744 scope.go:117] "RemoveContainer" containerID="0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.524933 4744 scope.go:117] "RemoveContainer" containerID="578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.554654 4744 scope.go:117] "RemoveContainer" containerID="0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5" Jan 06 14:38:59 crc kubenswrapper[4744]: E0106 14:38:59.556610 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5\": container with ID starting with 0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5 not found: ID does not exist" containerID="0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.556685 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5"} err="failed to get container status \"0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5\": rpc error: code = NotFound desc = could not find container \"0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5\": container with ID starting with 0db80f0044293014917e8c4b5114007243e41011ff29399e9366dd663f3d12c5 not found: ID does not exist" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.556734 4744 scope.go:117] "RemoveContainer" containerID="80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776" Jan 06 14:38:59 crc kubenswrapper[4744]: E0106 14:38:59.557393 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776\": container with ID starting with 80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776 not found: ID does not exist" containerID="80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.557474 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776"} err="failed to get container status \"80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776\": rpc error: code = NotFound desc = could not find container \"80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776\": container with ID starting with 80d7e4d1c76e2a350badd888a207620437da0ffc7b4a317b5cea5ec5a2ab8776 not found: ID does not exist" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.557528 4744 scope.go:117] "RemoveContainer" containerID="6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885" Jan 06 14:38:59 crc kubenswrapper[4744]: E0106 14:38:59.557915 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885\": container with ID starting with 6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885 not found: ID does not exist" containerID="6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.557937 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885"} err="failed to get container status \"6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885\": rpc error: code = NotFound desc = could not find container \"6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885\": container with ID starting with 6a9bc5f1d862c548f0e1299d320ded1355559004a542e5f892b1969a8df7e885 not found: ID does not exist" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.557954 4744 scope.go:117] "RemoveContainer" containerID="69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1" Jan 06 14:38:59 crc kubenswrapper[4744]: E0106 14:38:59.558352 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1\": container with ID starting with 69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1 not found: ID does not exist" containerID="69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.558371 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1"} err="failed to get container status \"69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1\": rpc error: code = NotFound desc = could not find container \"69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1\": container with ID starting with 69820a44e4fefdaf54f5a79c9824de5a846f3e4de983213eaba05366b5d4c3d1 not found: ID does not exist" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.558384 4744 scope.go:117] "RemoveContainer" containerID="0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a" Jan 06 14:38:59 crc kubenswrapper[4744]: E0106 14:38:59.558708 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a\": container with ID starting with 0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a not found: ID does not exist" containerID="0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.558724 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a"} err="failed to get container status \"0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a\": rpc error: code = NotFound desc = could not find container \"0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a\": container with ID starting with 0905b161e8339579c8a4c6528b8b46d2e364aab83170e5676c5101830e7d370a not found: ID does not exist" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.558734 4744 scope.go:117] "RemoveContainer" containerID="578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371" Jan 06 14:38:59 crc kubenswrapper[4744]: E0106 14:38:59.559148 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\": container with ID starting with 578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371 not found: ID does not exist" containerID="578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.559174 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371"} err="failed to get container status \"578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\": rpc error: code = NotFound desc = could not find container \"578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371\": container with ID starting with 578765e9e5a52b2250b6581f0e18de15f0b4971ceea4a0f9b7c0d177c3c16371 not found: ID does not exist" Jan 06 14:38:59 crc kubenswrapper[4744]: I0106 14:38:59.724143 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 06 14:39:03 crc kubenswrapper[4744]: I0106 14:39:03.714247 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:03 crc kubenswrapper[4744]: I0106 14:39:03.715750 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:07 crc kubenswrapper[4744]: E0106 14:39:07.049135 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:07 crc kubenswrapper[4744]: E0106 14:39:07.049797 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:07 crc kubenswrapper[4744]: E0106 14:39:07.050040 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:07 crc kubenswrapper[4744]: E0106 14:39:07.050393 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:07 crc kubenswrapper[4744]: E0106 14:39:07.050841 4744 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:07 crc kubenswrapper[4744]: I0106 14:39:07.050880 4744 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 06 14:39:07 crc kubenswrapper[4744]: E0106 14:39:07.051122 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" interval="200ms" Jan 06 14:39:07 crc kubenswrapper[4744]: E0106 14:39:07.251902 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" interval="400ms" Jan 06 14:39:07 crc kubenswrapper[4744]: E0106 14:39:07.390536 4744 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.22:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18882b2f970da538 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-06 14:38:57.14063084 +0000 UTC m=+133.768097178,LastTimestamp:2026-01-06 14:38:57.14063084 +0000 UTC m=+133.768097178,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 06 14:39:07 crc kubenswrapper[4744]: E0106 14:39:07.653346 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" interval="800ms" Jan 06 14:39:08 crc kubenswrapper[4744]: E0106 14:39:08.454777 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" interval="1.6s" Jan 06 14:39:10 crc kubenswrapper[4744]: E0106 14:39:10.056329 4744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.22:6443: connect: connection refused" interval="3.2s" Jan 06 14:39:10 crc kubenswrapper[4744]: I0106 14:39:10.710967 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:10 crc kubenswrapper[4744]: I0106 14:39:10.711828 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:10 crc kubenswrapper[4744]: I0106 14:39:10.712507 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:10 crc kubenswrapper[4744]: I0106 14:39:10.737988 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:10 crc kubenswrapper[4744]: I0106 14:39:10.738036 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:10 crc kubenswrapper[4744]: E0106 14:39:10.738637 4744 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:10 crc kubenswrapper[4744]: I0106 14:39:10.739009 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:10 crc kubenswrapper[4744]: W0106 14:39:10.771725 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-13439464d5e86d1d382db9a837f87e960921f3eea5e4d63bb1496fbfcdaba39f WatchSource:0}: Error finding container 13439464d5e86d1d382db9a837f87e960921f3eea5e4d63bb1496fbfcdaba39f: Status 404 returned error can't find the container with id 13439464d5e86d1d382db9a837f87e960921f3eea5e4d63bb1496fbfcdaba39f Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.488406 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.488821 4744 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a50ee7ce34ab8458b2d63585d172bfa39cc97f974fd6144e403ca0b36fc1be53" exitCode=1 Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.488939 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a50ee7ce34ab8458b2d63585d172bfa39cc97f974fd6144e403ca0b36fc1be53"} Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.489924 4744 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.489938 4744 scope.go:117] "RemoveContainer" containerID="a50ee7ce34ab8458b2d63585d172bfa39cc97f974fd6144e403ca0b36fc1be53" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.490436 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.490891 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.491118 4744 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="dd430adcf70e10dabee5fe02c6c34c0af7757b0782c49076a6463e9d6fc22595" exitCode=0 Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.491207 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"dd430adcf70e10dabee5fe02c6c34c0af7757b0782c49076a6463e9d6fc22595"} Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.491257 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"13439464d5e86d1d382db9a837f87e960921f3eea5e4d63bb1496fbfcdaba39f"} Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.491682 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.491704 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.491899 4744 status_manager.go:851] "Failed to get status for pod" podUID="c7684584-6c37-4d30-8593-70d738b41d93" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.492139 4744 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:11 crc kubenswrapper[4744]: E0106 14:39:11.492235 4744 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.492378 4744 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.22:6443: connect: connection refused" Jan 06 14:39:11 crc kubenswrapper[4744]: I0106 14:39:11.498045 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:39:11 crc kubenswrapper[4744]: E0106 14:39:11.741852 4744 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.22:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" volumeName="registry-storage" Jan 06 14:39:12 crc kubenswrapper[4744]: I0106 14:39:12.504393 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e885aeee144099bb296c30b43af835c731cfd5d0fd67b6775e73f2ff69181124"} Jan 06 14:39:12 crc kubenswrapper[4744]: I0106 14:39:12.504729 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e0d57bda9f58b373ac76ed76f26eaad4ad54b499cc6360d9c65721748dc3609a"} Jan 06 14:39:12 crc kubenswrapper[4744]: I0106 14:39:12.504747 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7b8ae2bf0a55db0b53b9994e17895d2e884d8424e666916aa41a8ffbf8858c61"} Jan 06 14:39:12 crc kubenswrapper[4744]: I0106 14:39:12.513004 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 06 14:39:12 crc kubenswrapper[4744]: I0106 14:39:12.513061 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"054f2d19d5c036dfac4a14147cb04e81c8e482cca924661626310be067356184"} Jan 06 14:39:13 crc kubenswrapper[4744]: I0106 14:39:13.519787 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:13 crc kubenswrapper[4744]: I0106 14:39:13.520000 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:13 crc kubenswrapper[4744]: I0106 14:39:13.520198 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b991f5e6da51c113cd974a5307a9e0de63eb33c014702b1438c3dd8338617bf6"} Jan 06 14:39:13 crc kubenswrapper[4744]: I0106 14:39:13.520218 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"66a0dc2b3af493918386add93ca6903cb2df16c42722f5b0e88f0038d0b0bb3f"} Jan 06 14:39:13 crc kubenswrapper[4744]: I0106 14:39:13.520243 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:14 crc kubenswrapper[4744]: I0106 14:39:14.423601 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:39:14 crc kubenswrapper[4744]: I0106 14:39:14.423662 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:39:15 crc kubenswrapper[4744]: I0106 14:39:15.740117 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:15 crc kubenswrapper[4744]: I0106 14:39:15.740478 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:15 crc kubenswrapper[4744]: I0106 14:39:15.746543 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:16 crc kubenswrapper[4744]: I0106 14:39:16.051947 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:39:16 crc kubenswrapper[4744]: I0106 14:39:16.052299 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 06 14:39:16 crc kubenswrapper[4744]: I0106 14:39:16.052461 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 06 14:39:18 crc kubenswrapper[4744]: I0106 14:39:18.598762 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:39:18 crc kubenswrapper[4744]: I0106 14:39:18.800668 4744 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:19 crc kubenswrapper[4744]: I0106 14:39:19.557567 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:19 crc kubenswrapper[4744]: I0106 14:39:19.557616 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:19 crc kubenswrapper[4744]: I0106 14:39:19.565067 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:19 crc kubenswrapper[4744]: I0106 14:39:19.568835 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="12623660-58da-425a-8826-f76a4afa5400" Jan 06 14:39:20 crc kubenswrapper[4744]: I0106 14:39:20.564479 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:20 crc kubenswrapper[4744]: I0106 14:39:20.564868 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:23 crc kubenswrapper[4744]: I0106 14:39:23.742235 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="12623660-58da-425a-8826-f76a4afa5400" Jan 06 14:39:26 crc kubenswrapper[4744]: I0106 14:39:26.052733 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 06 14:39:26 crc kubenswrapper[4744]: I0106 14:39:26.053068 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 06 14:39:28 crc kubenswrapper[4744]: I0106 14:39:28.554780 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 06 14:39:28 crc kubenswrapper[4744]: I0106 14:39:28.565367 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 06 14:39:28 crc kubenswrapper[4744]: I0106 14:39:28.858908 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 06 14:39:28 crc kubenswrapper[4744]: I0106 14:39:28.929418 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 06 14:39:29 crc kubenswrapper[4744]: I0106 14:39:29.284020 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 06 14:39:29 crc kubenswrapper[4744]: I0106 14:39:29.300090 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 06 14:39:29 crc kubenswrapper[4744]: I0106 14:39:29.330337 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 06 14:39:29 crc kubenswrapper[4744]: I0106 14:39:29.428666 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 06 14:39:29 crc kubenswrapper[4744]: I0106 14:39:29.762109 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 06 14:39:30 crc kubenswrapper[4744]: I0106 14:39:30.285578 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 06 14:39:30 crc kubenswrapper[4744]: I0106 14:39:30.551818 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 06 14:39:30 crc kubenswrapper[4744]: I0106 14:39:30.816578 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.036979 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.052265 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.418521 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.454069 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.481972 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.509107 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.676795 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.711905 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.791024 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 06 14:39:31 crc kubenswrapper[4744]: I0106 14:39:31.823175 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.006669 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.262318 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.268432 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.395356 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.427874 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.457232 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.478877 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.569995 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.573803 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.597681 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.759534 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 06 14:39:32 crc kubenswrapper[4744]: I0106 14:39:32.805950 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.056810 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.110709 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.255463 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.470756 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.474853 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.512994 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.571251 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.715488 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.719071 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.865459 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.949026 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 06 14:39:33 crc kubenswrapper[4744]: I0106 14:39:33.987520 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.141504 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.194264 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.238497 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.313421 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.343811 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.418695 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.456019 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.485681 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.495089 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.565206 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.587440 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.648401 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.711917 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.758659 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.766375 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 06 14:39:34 crc kubenswrapper[4744]: I0106 14:39:34.840049 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.144931 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.150717 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.183611 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.197846 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.199652 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.200090 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.220038 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.294825 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.311094 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.387352 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.443140 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.466776 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.517373 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.520282 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.580220 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.592323 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.615441 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.739858 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.896514 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.898808 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 06 14:39:35 crc kubenswrapper[4744]: I0106 14:39:35.991096 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.051665 4744 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.051731 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.051790 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.052474 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"054f2d19d5c036dfac4a14147cb04e81c8e482cca924661626310be067356184"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.052648 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://054f2d19d5c036dfac4a14147cb04e81c8e482cca924661626310be067356184" gracePeriod=30 Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.098299 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.123231 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.189364 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.209306 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.228598 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.245473 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.280585 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.280920 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.286211 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.286425 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.401816 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.444891 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.466032 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.474754 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.482455 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.519127 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.530466 4744 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.562858 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.571554 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.571943 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.628248 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.628492 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.701562 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.739448 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.757018 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.771141 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.808702 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.828968 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.941968 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.944667 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 06 14:39:36 crc kubenswrapper[4744]: I0106 14:39:36.951585 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.026481 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.057684 4744 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.060694 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=41.0606605 podStartE2EDuration="41.0606605s" podCreationTimestamp="2026-01-06 14:38:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:39:18.286473741 +0000 UTC m=+154.913940069" watchObservedRunningTime="2026-01-06 14:39:37.0606605 +0000 UTC m=+173.688126868" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.068444 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.068542 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-75b7b58d79-qlsgj"] Jan 06 14:39:37 crc kubenswrapper[4744]: E0106 14:39:37.068934 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7684584-6c37-4d30-8593-70d738b41d93" containerName="installer" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.068973 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7684584-6c37-4d30-8593-70d738b41d93" containerName="installer" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.069027 4744 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.069073 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="85d85475-e337-4034-8a0c-1f6e873d9476" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.069248 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7684584-6c37-4d30-8593-70d738b41d93" containerName="installer" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.070211 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.081173 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.082407 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.082789 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.083072 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.083126 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.083615 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.083900 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.084049 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.084665 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.087063 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.087385 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.092088 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.093714 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.098826 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.099298 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.099985 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-audit-dir\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100027 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100062 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-service-ca\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-audit-policies\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100116 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89qjt\" (UniqueName: \"kubernetes.io/projected/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-kube-api-access-89qjt\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100243 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-template-login\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100292 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-template-error\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100320 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-session\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100354 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100385 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100427 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100464 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100504 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.100535 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-router-certs\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.101124 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.103586 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.108265 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.128317 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.128303072 podStartE2EDuration="19.128303072s" podCreationTimestamp="2026-01-06 14:39:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:39:37.12408042 +0000 UTC m=+173.751546758" watchObservedRunningTime="2026-01-06 14:39:37.128303072 +0000 UTC m=+173.755769390" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.201433 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-audit-dir\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203287 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.201608 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-audit-dir\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203349 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-service-ca\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203374 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-audit-policies\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203402 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89qjt\" (UniqueName: \"kubernetes.io/projected/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-kube-api-access-89qjt\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-template-login\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203463 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-template-error\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203481 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-session\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203506 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203545 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203583 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203614 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203648 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.203675 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-router-certs\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.204856 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.205358 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-audit-policies\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.205845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-service-ca\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.206133 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.209683 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-session\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.209754 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-router-certs\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.209935 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.210918 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.212337 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-template-login\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.215280 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-user-template-error\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.215746 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.216055 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.221465 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.223701 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89qjt\" (UniqueName: \"kubernetes.io/projected/22ed66ba-2d7d-464f-ab3c-dd20b773f1ca-kube-api-access-89qjt\") pod \"oauth-openshift-75b7b58d79-qlsgj\" (UID: \"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca\") " pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.264887 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.322537 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.330931 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.344537 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.372706 4744 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.400375 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.428855 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.449348 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.475539 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.479330 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.551446 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.648333 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.649483 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.743905 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.788363 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.946085 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 06 14:39:37 crc kubenswrapper[4744]: I0106 14:39:37.969950 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.004495 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.018479 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-75b7b58d79-qlsgj"] Jan 06 14:39:38 crc kubenswrapper[4744]: W0106 14:39:38.028148 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22ed66ba_2d7d_464f_ab3c_dd20b773f1ca.slice/crio-4297fe31e9e024e57d0bf385f4f707af6918c5a4eaf962c5268eb53b54fdda14 WatchSource:0}: Error finding container 4297fe31e9e024e57d0bf385f4f707af6918c5a4eaf962c5268eb53b54fdda14: Status 404 returned error can't find the container with id 4297fe31e9e024e57d0bf385f4f707af6918c5a4eaf962c5268eb53b54fdda14 Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.067744 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.260314 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.347446 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.349795 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.356705 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.358663 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.410861 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.423508 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.446930 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.630367 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.660942 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.667239 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.672638 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.693981 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" event={"ID":"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca","Type":"ContainerStarted","Data":"654f12240d0fc35fc9b81bda51042667ef19fddb087396c547669e4d302bffaa"} Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.694051 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" event={"ID":"22ed66ba-2d7d-464f-ab3c-dd20b773f1ca","Type":"ContainerStarted","Data":"4297fe31e9e024e57d0bf385f4f707af6918c5a4eaf962c5268eb53b54fdda14"} Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.694477 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.702674 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.740529 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-75b7b58d79-qlsgj" podStartSLOduration=75.740501768 podStartE2EDuration="1m15.740501768s" podCreationTimestamp="2026-01-06 14:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:39:38.737727014 +0000 UTC m=+175.365193332" watchObservedRunningTime="2026-01-06 14:39:38.740501768 +0000 UTC m=+175.367968096" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.768216 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.768280 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.811574 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.820657 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.862085 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.899240 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.941578 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.958352 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 06 14:39:38 crc kubenswrapper[4744]: I0106 14:39:38.985016 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.023127 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.042829 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.227623 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.228903 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.256991 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.297068 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.391762 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.439827 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.447456 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.512767 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.558711 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.573218 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.701042 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.760428 4744 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.824608 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.826118 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.864901 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 06 14:39:39 crc kubenswrapper[4744]: I0106 14:39:39.931852 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.060708 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.204466 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.263820 4744 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.316749 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.318991 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.352771 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.383041 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.408029 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.412934 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.486466 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.515803 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.749071 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.761568 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.958952 4744 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.959337 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f" gracePeriod=5 Jan 06 14:39:40 crc kubenswrapper[4744]: I0106 14:39:40.966390 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.032049 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.071140 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.093757 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.105434 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.130039 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.209996 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.212527 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.400042 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.438328 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.476506 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.518834 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.609420 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.648069 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.690192 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.747024 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.842769 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.954365 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 06 14:39:41 crc kubenswrapper[4744]: I0106 14:39:41.972923 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.023062 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.090422 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.091687 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.243482 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.295920 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.302610 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.420047 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.628928 4744 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.703812 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.797238 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.891241 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 06 14:39:42 crc kubenswrapper[4744]: I0106 14:39:42.954963 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 06 14:39:43 crc kubenswrapper[4744]: I0106 14:39:43.108206 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 06 14:39:43 crc kubenswrapper[4744]: I0106 14:39:43.182825 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 06 14:39:43 crc kubenswrapper[4744]: I0106 14:39:43.197298 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 06 14:39:43 crc kubenswrapper[4744]: I0106 14:39:43.315257 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 06 14:39:43 crc kubenswrapper[4744]: I0106 14:39:43.364466 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 06 14:39:43 crc kubenswrapper[4744]: I0106 14:39:43.492391 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 06 14:39:43 crc kubenswrapper[4744]: I0106 14:39:43.741720 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 06 14:39:43 crc kubenswrapper[4744]: I0106 14:39:43.867245 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 06 14:39:43 crc kubenswrapper[4744]: I0106 14:39:43.917988 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 06 14:39:44 crc kubenswrapper[4744]: I0106 14:39:44.124532 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 06 14:39:44 crc kubenswrapper[4744]: I0106 14:39:44.304951 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 06 14:39:44 crc kubenswrapper[4744]: I0106 14:39:44.312631 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 06 14:39:44 crc kubenswrapper[4744]: I0106 14:39:44.424215 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:39:44 crc kubenswrapper[4744]: I0106 14:39:44.424320 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:39:44 crc kubenswrapper[4744]: I0106 14:39:44.745366 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 06 14:39:44 crc kubenswrapper[4744]: I0106 14:39:44.828422 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 06 14:39:45 crc kubenswrapper[4744]: I0106 14:39:45.840222 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.573015 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.573130 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.681433 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.681523 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.681575 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.681598 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.681649 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.681749 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.681804 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.681743 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.682007 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.682518 4744 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.682548 4744 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.682566 4744 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.682583 4744 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.696068 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.753030 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.753111 4744 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f" exitCode=137 Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.753195 4744 scope.go:117] "RemoveContainer" containerID="73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.753402 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.780374 4744 scope.go:117] "RemoveContainer" containerID="73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f" Jan 06 14:39:46 crc kubenswrapper[4744]: E0106 14:39:46.780978 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f\": container with ID starting with 73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f not found: ID does not exist" containerID="73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.781065 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f"} err="failed to get container status \"73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f\": rpc error: code = NotFound desc = could not find container \"73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f\": container with ID starting with 73cc1ecdb7f9ce571f0b30f6e35b4f799447a76e2ada5d914ffacdfacc6b9e4f not found: ID does not exist" Jan 06 14:39:46 crc kubenswrapper[4744]: I0106 14:39:46.784382 4744 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 06 14:39:47 crc kubenswrapper[4744]: I0106 14:39:47.722446 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 06 14:39:47 crc kubenswrapper[4744]: I0106 14:39:47.722927 4744 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 06 14:39:47 crc kubenswrapper[4744]: I0106 14:39:47.737484 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 06 14:39:47 crc kubenswrapper[4744]: I0106 14:39:47.737737 4744 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="67518773-3479-4357-b474-d1bfc9a51075" Jan 06 14:39:47 crc kubenswrapper[4744]: I0106 14:39:47.744379 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 06 14:39:47 crc kubenswrapper[4744]: I0106 14:39:47.744439 4744 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="67518773-3479-4357-b474-d1bfc9a51075" Jan 06 14:39:55 crc kubenswrapper[4744]: I0106 14:39:55.763874 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 06 14:40:06 crc kubenswrapper[4744]: I0106 14:40:06.917811 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 06 14:40:06 crc kubenswrapper[4744]: I0106 14:40:06.920253 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 06 14:40:06 crc kubenswrapper[4744]: I0106 14:40:06.920306 4744 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="054f2d19d5c036dfac4a14147cb04e81c8e482cca924661626310be067356184" exitCode=137 Jan 06 14:40:06 crc kubenswrapper[4744]: I0106 14:40:06.920349 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"054f2d19d5c036dfac4a14147cb04e81c8e482cca924661626310be067356184"} Jan 06 14:40:06 crc kubenswrapper[4744]: I0106 14:40:06.920415 4744 scope.go:117] "RemoveContainer" containerID="a50ee7ce34ab8458b2d63585d172bfa39cc97f974fd6144e403ca0b36fc1be53" Jan 06 14:40:07 crc kubenswrapper[4744]: I0106 14:40:07.928635 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 06 14:40:07 crc kubenswrapper[4744]: I0106 14:40:07.930314 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"20ae756bc6e1c5bee27b0b6e489d1b101dc4fcf825828b0614d6f11233ba282d"} Jan 06 14:40:08 crc kubenswrapper[4744]: I0106 14:40:08.599278 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:40:14 crc kubenswrapper[4744]: I0106 14:40:14.424084 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:40:14 crc kubenswrapper[4744]: I0106 14:40:14.424264 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:40:14 crc kubenswrapper[4744]: I0106 14:40:14.424360 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:40:14 crc kubenswrapper[4744]: I0106 14:40:14.425447 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f0db91145c265a1792dad8ccd75d8c9a1942a698f2bf63e4a4bdd7ae0b74f422"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 14:40:14 crc kubenswrapper[4744]: I0106 14:40:14.425563 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://f0db91145c265a1792dad8ccd75d8c9a1942a698f2bf63e4a4bdd7ae0b74f422" gracePeriod=600 Jan 06 14:40:14 crc kubenswrapper[4744]: I0106 14:40:14.986822 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="f0db91145c265a1792dad8ccd75d8c9a1942a698f2bf63e4a4bdd7ae0b74f422" exitCode=0 Jan 06 14:40:14 crc kubenswrapper[4744]: I0106 14:40:14.986909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"f0db91145c265a1792dad8ccd75d8c9a1942a698f2bf63e4a4bdd7ae0b74f422"} Jan 06 14:40:14 crc kubenswrapper[4744]: I0106 14:40:14.987112 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"a62092cf79e8012550863bd594759780ea4401103ffbb749d1491e8c21a56e8a"} Jan 06 14:40:16 crc kubenswrapper[4744]: I0106 14:40:16.051621 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:40:16 crc kubenswrapper[4744]: I0106 14:40:16.055496 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:40:17 crc kubenswrapper[4744]: I0106 14:40:17.585238 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 06 14:40:18 crc kubenswrapper[4744]: I0106 14:40:18.603996 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.402014 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b68d4bc88-v86n5"] Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.402782 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" podUID="360f121c-7279-4311-9331-3abf4d0dd528" containerName="controller-manager" containerID="cri-o://ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888" gracePeriod=30 Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.408835 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl"] Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.409605 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" podUID="7046fa2e-834b-4802-a76c-d70de3855872" containerName="route-controller-manager" containerID="cri-o://8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a" gracePeriod=30 Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.833706 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.840900 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.961512 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-client-ca\") pod \"7046fa2e-834b-4802-a76c-d70de3855872\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.961554 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75fj5\" (UniqueName: \"kubernetes.io/projected/7046fa2e-834b-4802-a76c-d70de3855872-kube-api-access-75fj5\") pod \"7046fa2e-834b-4802-a76c-d70de3855872\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.961624 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7046fa2e-834b-4802-a76c-d70de3855872-serving-cert\") pod \"7046fa2e-834b-4802-a76c-d70de3855872\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.961645 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-proxy-ca-bundles\") pod \"360f121c-7279-4311-9331-3abf4d0dd528\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.961670 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/360f121c-7279-4311-9331-3abf4d0dd528-serving-cert\") pod \"360f121c-7279-4311-9331-3abf4d0dd528\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.961693 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-config\") pod \"7046fa2e-834b-4802-a76c-d70de3855872\" (UID: \"7046fa2e-834b-4802-a76c-d70de3855872\") " Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.962613 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "360f121c-7279-4311-9331-3abf4d0dd528" (UID: "360f121c-7279-4311-9331-3abf4d0dd528"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.962615 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-client-ca" (OuterVolumeSpecName: "client-ca") pod "7046fa2e-834b-4802-a76c-d70de3855872" (UID: "7046fa2e-834b-4802-a76c-d70de3855872"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.962706 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-config\") pod \"360f121c-7279-4311-9331-3abf4d0dd528\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.962829 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-config" (OuterVolumeSpecName: "config") pod "7046fa2e-834b-4802-a76c-d70de3855872" (UID: "7046fa2e-834b-4802-a76c-d70de3855872"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.963366 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-client-ca" (OuterVolumeSpecName: "client-ca") pod "360f121c-7279-4311-9331-3abf4d0dd528" (UID: "360f121c-7279-4311-9331-3abf4d0dd528"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.963380 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-config" (OuterVolumeSpecName: "config") pod "360f121c-7279-4311-9331-3abf4d0dd528" (UID: "360f121c-7279-4311-9331-3abf4d0dd528"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.962739 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-client-ca\") pod \"360f121c-7279-4311-9331-3abf4d0dd528\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.963452 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfg5c\" (UniqueName: \"kubernetes.io/projected/360f121c-7279-4311-9331-3abf4d0dd528-kube-api-access-hfg5c\") pod \"360f121c-7279-4311-9331-3abf4d0dd528\" (UID: \"360f121c-7279-4311-9331-3abf4d0dd528\") " Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.963681 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.963700 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.963709 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.963716 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7046fa2e-834b-4802-a76c-d70de3855872-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.963724 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/360f121c-7279-4311-9331-3abf4d0dd528-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.966893 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7046fa2e-834b-4802-a76c-d70de3855872-kube-api-access-75fj5" (OuterVolumeSpecName: "kube-api-access-75fj5") pod "7046fa2e-834b-4802-a76c-d70de3855872" (UID: "7046fa2e-834b-4802-a76c-d70de3855872"). InnerVolumeSpecName "kube-api-access-75fj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.966941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/360f121c-7279-4311-9331-3abf4d0dd528-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "360f121c-7279-4311-9331-3abf4d0dd528" (UID: "360f121c-7279-4311-9331-3abf4d0dd528"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.970493 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7046fa2e-834b-4802-a76c-d70de3855872-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7046fa2e-834b-4802-a76c-d70de3855872" (UID: "7046fa2e-834b-4802-a76c-d70de3855872"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:40:24 crc kubenswrapper[4744]: I0106 14:40:24.970952 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/360f121c-7279-4311-9331-3abf4d0dd528-kube-api-access-hfg5c" (OuterVolumeSpecName: "kube-api-access-hfg5c") pod "360f121c-7279-4311-9331-3abf4d0dd528" (UID: "360f121c-7279-4311-9331-3abf4d0dd528"). InnerVolumeSpecName "kube-api-access-hfg5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.050075 4744 generic.go:334] "Generic (PLEG): container finished" podID="360f121c-7279-4311-9331-3abf4d0dd528" containerID="ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888" exitCode=0 Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.050192 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" event={"ID":"360f121c-7279-4311-9331-3abf4d0dd528","Type":"ContainerDied","Data":"ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888"} Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.050204 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.050237 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b68d4bc88-v86n5" event={"ID":"360f121c-7279-4311-9331-3abf4d0dd528","Type":"ContainerDied","Data":"e1536443832322e9c236dcdf1a6ca0fed98181ffe223b5245b786331971ab139"} Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.050269 4744 scope.go:117] "RemoveContainer" containerID="ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.052728 4744 generic.go:334] "Generic (PLEG): container finished" podID="7046fa2e-834b-4802-a76c-d70de3855872" containerID="8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a" exitCode=0 Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.052792 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.052800 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" event={"ID":"7046fa2e-834b-4802-a76c-d70de3855872","Type":"ContainerDied","Data":"8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a"} Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.052862 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl" event={"ID":"7046fa2e-834b-4802-a76c-d70de3855872","Type":"ContainerDied","Data":"8c84fd0011397647c33b7b32c28d6b89e808407fb93abfe6ed92ef8d5a24ecd3"} Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.065101 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7046fa2e-834b-4802-a76c-d70de3855872-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.065489 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/360f121c-7279-4311-9331-3abf4d0dd528-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.065503 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfg5c\" (UniqueName: \"kubernetes.io/projected/360f121c-7279-4311-9331-3abf4d0dd528-kube-api-access-hfg5c\") on node \"crc\" DevicePath \"\"" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.065515 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75fj5\" (UniqueName: \"kubernetes.io/projected/7046fa2e-834b-4802-a76c-d70de3855872-kube-api-access-75fj5\") on node \"crc\" DevicePath \"\"" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.083591 4744 scope.go:117] "RemoveContainer" containerID="ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888" Jan 06 14:40:25 crc kubenswrapper[4744]: E0106 14:40:25.086431 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888\": container with ID starting with ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888 not found: ID does not exist" containerID="ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.086489 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888"} err="failed to get container status \"ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888\": rpc error: code = NotFound desc = could not find container \"ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888\": container with ID starting with ce5ee1aae5267aa89990897fd3afd5616357e36216c29ef3a792f58a73b70888 not found: ID does not exist" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.086526 4744 scope.go:117] "RemoveContainer" containerID="8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.091607 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl"] Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.101504 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6f5bb76d5c-slktl"] Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.111884 4744 scope.go:117] "RemoveContainer" containerID="8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a" Jan 06 14:40:25 crc kubenswrapper[4744]: E0106 14:40:25.112483 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a\": container with ID starting with 8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a not found: ID does not exist" containerID="8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.112531 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a"} err="failed to get container status \"8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a\": rpc error: code = NotFound desc = could not find container \"8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a\": container with ID starting with 8007873e977067eaa6200a31d8a140cb276565dd444129e5d16ae2ed2559599a not found: ID does not exist" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.136430 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b68d4bc88-v86n5"] Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.149203 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6b68d4bc88-v86n5"] Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.721389 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="360f121c-7279-4311-9331-3abf4d0dd528" path="/var/lib/kubelet/pods/360f121c-7279-4311-9331-3abf4d0dd528/volumes" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.722781 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7046fa2e-834b-4802-a76c-d70de3855872" path="/var/lib/kubelet/pods/7046fa2e-834b-4802-a76c-d70de3855872/volumes" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.983268 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66f55d8b9f-wsn84"] Jan 06 14:40:25 crc kubenswrapper[4744]: E0106 14:40:25.983511 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7046fa2e-834b-4802-a76c-d70de3855872" containerName="route-controller-manager" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.983527 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7046fa2e-834b-4802-a76c-d70de3855872" containerName="route-controller-manager" Jan 06 14:40:25 crc kubenswrapper[4744]: E0106 14:40:25.983539 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="360f121c-7279-4311-9331-3abf4d0dd528" containerName="controller-manager" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.983548 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="360f121c-7279-4311-9331-3abf4d0dd528" containerName="controller-manager" Jan 06 14:40:25 crc kubenswrapper[4744]: E0106 14:40:25.983565 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.983574 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.983693 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7046fa2e-834b-4802-a76c-d70de3855872" containerName="route-controller-manager" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.983727 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="360f121c-7279-4311-9331-3abf4d0dd528" containerName="controller-manager" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.983739 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.984243 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.987397 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx"] Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.988043 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.988896 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.988911 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.988911 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.989940 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.990227 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.993119 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.995326 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.997485 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.997697 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.997916 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 06 14:40:25 crc kubenswrapper[4744]: I0106 14:40:25.998334 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.000683 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.002032 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f55d8b9f-wsn84"] Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.008618 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.012105 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx"] Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.180905 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66691427-c48c-4dbe-b5f1-e512b5d33ae6-serving-cert\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.181018 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsbhg\" (UniqueName: \"kubernetes.io/projected/66691427-c48c-4dbe-b5f1-e512b5d33ae6-kube-api-access-vsbhg\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.181360 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-client-ca\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.181510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-config\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.181603 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gpp2\" (UniqueName: \"kubernetes.io/projected/a07b6c09-179a-4562-ac56-5ac471127525-kube-api-access-2gpp2\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.181671 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a07b6c09-179a-4562-ac56-5ac471127525-serving-cert\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.181721 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-proxy-ca-bundles\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.181843 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-config\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.181920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-client-ca\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.282687 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-config\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.282761 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-client-ca\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.282784 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66691427-c48c-4dbe-b5f1-e512b5d33ae6-serving-cert\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.282807 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsbhg\" (UniqueName: \"kubernetes.io/projected/66691427-c48c-4dbe-b5f1-e512b5d33ae6-kube-api-access-vsbhg\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.282851 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-client-ca\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.282880 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-config\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.282899 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gpp2\" (UniqueName: \"kubernetes.io/projected/a07b6c09-179a-4562-ac56-5ac471127525-kube-api-access-2gpp2\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.282927 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a07b6c09-179a-4562-ac56-5ac471127525-serving-cert\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.282948 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-proxy-ca-bundles\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.284323 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-client-ca\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.284335 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-proxy-ca-bundles\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.284452 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-client-ca\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.285027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-config\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.287042 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-config\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.289884 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a07b6c09-179a-4562-ac56-5ac471127525-serving-cert\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.290984 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66691427-c48c-4dbe-b5f1-e512b5d33ae6-serving-cert\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.303803 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsbhg\" (UniqueName: \"kubernetes.io/projected/66691427-c48c-4dbe-b5f1-e512b5d33ae6-kube-api-access-vsbhg\") pod \"controller-manager-66f55d8b9f-wsn84\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.307533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gpp2\" (UniqueName: \"kubernetes.io/projected/a07b6c09-179a-4562-ac56-5ac471127525-kube-api-access-2gpp2\") pod \"route-controller-manager-fd6f6d45-7kcmx\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.309435 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.328100 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.623671 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx"] Jan 06 14:40:26 crc kubenswrapper[4744]: W0106 14:40:26.635468 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda07b6c09_179a_4562_ac56_5ac471127525.slice/crio-4a308e65c73e4c9ae675b1de75d576b34239885dd3a1a6138e1d34b237fd91e3 WatchSource:0}: Error finding container 4a308e65c73e4c9ae675b1de75d576b34239885dd3a1a6138e1d34b237fd91e3: Status 404 returned error can't find the container with id 4a308e65c73e4c9ae675b1de75d576b34239885dd3a1a6138e1d34b237fd91e3 Jan 06 14:40:26 crc kubenswrapper[4744]: I0106 14:40:26.777804 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f55d8b9f-wsn84"] Jan 06 14:40:26 crc kubenswrapper[4744]: W0106 14:40:26.784931 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66691427_c48c_4dbe_b5f1_e512b5d33ae6.slice/crio-ae46b6e36bcc877a17a78f96d9934f7691d5276b7a18c42069fad30bf6725f35 WatchSource:0}: Error finding container ae46b6e36bcc877a17a78f96d9934f7691d5276b7a18c42069fad30bf6725f35: Status 404 returned error can't find the container with id ae46b6e36bcc877a17a78f96d9934f7691d5276b7a18c42069fad30bf6725f35 Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.069943 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" event={"ID":"66691427-c48c-4dbe-b5f1-e512b5d33ae6","Type":"ContainerStarted","Data":"5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368"} Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.070257 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.070271 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" event={"ID":"66691427-c48c-4dbe-b5f1-e512b5d33ae6","Type":"ContainerStarted","Data":"ae46b6e36bcc877a17a78f96d9934f7691d5276b7a18c42069fad30bf6725f35"} Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.072828 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" event={"ID":"a07b6c09-179a-4562-ac56-5ac471127525","Type":"ContainerStarted","Data":"3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9"} Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.072873 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" event={"ID":"a07b6c09-179a-4562-ac56-5ac471127525","Type":"ContainerStarted","Data":"4a308e65c73e4c9ae675b1de75d576b34239885dd3a1a6138e1d34b237fd91e3"} Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.073075 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.075421 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.090053 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" podStartSLOduration=3.090029162 podStartE2EDuration="3.090029162s" podCreationTimestamp="2026-01-06 14:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:40:27.087928365 +0000 UTC m=+223.715394683" watchObservedRunningTime="2026-01-06 14:40:27.090029162 +0000 UTC m=+223.717495480" Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.111179 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" podStartSLOduration=3.11113018 podStartE2EDuration="3.11113018s" podCreationTimestamp="2026-01-06 14:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:40:27.107675916 +0000 UTC m=+223.735142234" watchObservedRunningTime="2026-01-06 14:40:27.11113018 +0000 UTC m=+223.738596528" Jan 06 14:40:27 crc kubenswrapper[4744]: I0106 14:40:27.272664 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.273922 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cvggf"] Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.275464 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.298719 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cvggf"] Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.394994 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b40ab104-7596-4e6a-8a37-5ae516143646-registry-certificates\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.395514 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.396127 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b40ab104-7596-4e6a-8a37-5ae516143646-bound-sa-token\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.396266 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b40ab104-7596-4e6a-8a37-5ae516143646-trusted-ca\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.396391 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4cqw\" (UniqueName: \"kubernetes.io/projected/b40ab104-7596-4e6a-8a37-5ae516143646-kube-api-access-r4cqw\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.396514 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b40ab104-7596-4e6a-8a37-5ae516143646-registry-tls\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.396629 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b40ab104-7596-4e6a-8a37-5ae516143646-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.396752 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b40ab104-7596-4e6a-8a37-5ae516143646-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.414774 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.498510 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b40ab104-7596-4e6a-8a37-5ae516143646-trusted-ca\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.498584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4cqw\" (UniqueName: \"kubernetes.io/projected/b40ab104-7596-4e6a-8a37-5ae516143646-kube-api-access-r4cqw\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.498614 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b40ab104-7596-4e6a-8a37-5ae516143646-registry-tls\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.498638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b40ab104-7596-4e6a-8a37-5ae516143646-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.498677 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b40ab104-7596-4e6a-8a37-5ae516143646-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.498751 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b40ab104-7596-4e6a-8a37-5ae516143646-registry-certificates\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.498779 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b40ab104-7596-4e6a-8a37-5ae516143646-bound-sa-token\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.499689 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b40ab104-7596-4e6a-8a37-5ae516143646-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.500509 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b40ab104-7596-4e6a-8a37-5ae516143646-registry-certificates\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.500612 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b40ab104-7596-4e6a-8a37-5ae516143646-trusted-ca\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.505025 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b40ab104-7596-4e6a-8a37-5ae516143646-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.505185 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b40ab104-7596-4e6a-8a37-5ae516143646-registry-tls\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.558744 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b40ab104-7596-4e6a-8a37-5ae516143646-bound-sa-token\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.558888 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4cqw\" (UniqueName: \"kubernetes.io/projected/b40ab104-7596-4e6a-8a37-5ae516143646-kube-api-access-r4cqw\") pod \"image-registry-66df7c8f76-cvggf\" (UID: \"b40ab104-7596-4e6a-8a37-5ae516143646\") " pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.591064 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:52 crc kubenswrapper[4744]: I0106 14:40:52.960503 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cvggf"] Jan 06 14:40:53 crc kubenswrapper[4744]: I0106 14:40:53.227543 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" event={"ID":"b40ab104-7596-4e6a-8a37-5ae516143646","Type":"ContainerStarted","Data":"3fdda2ef86db7c9090505e0e85721885818fb875fce5fd149885540c928ae1e7"} Jan 06 14:40:53 crc kubenswrapper[4744]: I0106 14:40:53.227604 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" event={"ID":"b40ab104-7596-4e6a-8a37-5ae516143646","Type":"ContainerStarted","Data":"cc3a7e516ab81b5f438cf071b8930b3f116d176d41a8d50f8f616f0271c90177"} Jan 06 14:40:53 crc kubenswrapper[4744]: I0106 14:40:53.227694 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:40:53 crc kubenswrapper[4744]: I0106 14:40:53.248385 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" podStartSLOduration=1.248361314 podStartE2EDuration="1.248361314s" podCreationTimestamp="2026-01-06 14:40:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:40:53.247872341 +0000 UTC m=+249.875338669" watchObservedRunningTime="2026-01-06 14:40:53.248361314 +0000 UTC m=+249.875827632" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.183240 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66f55d8b9f-wsn84"] Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.184050 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" podUID="66691427-c48c-4dbe-b5f1-e512b5d33ae6" containerName="controller-manager" containerID="cri-o://5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368" gracePeriod=30 Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.266795 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx"] Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.267339 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" podUID="a07b6c09-179a-4562-ac56-5ac471127525" containerName="route-controller-manager" containerID="cri-o://3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9" gracePeriod=30 Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.630803 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.636594 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.773066 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-client-ca\") pod \"a07b6c09-179a-4562-ac56-5ac471127525\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.773114 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-proxy-ca-bundles\") pod \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.773133 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66691427-c48c-4dbe-b5f1-e512b5d33ae6-serving-cert\") pod \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.773190 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gpp2\" (UniqueName: \"kubernetes.io/projected/a07b6c09-179a-4562-ac56-5ac471127525-kube-api-access-2gpp2\") pod \"a07b6c09-179a-4562-ac56-5ac471127525\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.773214 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-client-ca\") pod \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.773233 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a07b6c09-179a-4562-ac56-5ac471127525-serving-cert\") pod \"a07b6c09-179a-4562-ac56-5ac471127525\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.773254 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsbhg\" (UniqueName: \"kubernetes.io/projected/66691427-c48c-4dbe-b5f1-e512b5d33ae6-kube-api-access-vsbhg\") pod \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.773275 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-config\") pod \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\" (UID: \"66691427-c48c-4dbe-b5f1-e512b5d33ae6\") " Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.773296 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-config\") pod \"a07b6c09-179a-4562-ac56-5ac471127525\" (UID: \"a07b6c09-179a-4562-ac56-5ac471127525\") " Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.774061 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-client-ca" (OuterVolumeSpecName: "client-ca") pod "a07b6c09-179a-4562-ac56-5ac471127525" (UID: "a07b6c09-179a-4562-ac56-5ac471127525"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.774595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-config" (OuterVolumeSpecName: "config") pod "a07b6c09-179a-4562-ac56-5ac471127525" (UID: "a07b6c09-179a-4562-ac56-5ac471127525"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.775303 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-config" (OuterVolumeSpecName: "config") pod "66691427-c48c-4dbe-b5f1-e512b5d33ae6" (UID: "66691427-c48c-4dbe-b5f1-e512b5d33ae6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.775392 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-client-ca" (OuterVolumeSpecName: "client-ca") pod "66691427-c48c-4dbe-b5f1-e512b5d33ae6" (UID: "66691427-c48c-4dbe-b5f1-e512b5d33ae6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.775667 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "66691427-c48c-4dbe-b5f1-e512b5d33ae6" (UID: "66691427-c48c-4dbe-b5f1-e512b5d33ae6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.780854 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66691427-c48c-4dbe-b5f1-e512b5d33ae6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "66691427-c48c-4dbe-b5f1-e512b5d33ae6" (UID: "66691427-c48c-4dbe-b5f1-e512b5d33ae6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.781291 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a07b6c09-179a-4562-ac56-5ac471127525-kube-api-access-2gpp2" (OuterVolumeSpecName: "kube-api-access-2gpp2") pod "a07b6c09-179a-4562-ac56-5ac471127525" (UID: "a07b6c09-179a-4562-ac56-5ac471127525"). InnerVolumeSpecName "kube-api-access-2gpp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.781527 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a07b6c09-179a-4562-ac56-5ac471127525-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a07b6c09-179a-4562-ac56-5ac471127525" (UID: "a07b6c09-179a-4562-ac56-5ac471127525"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.781674 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66691427-c48c-4dbe-b5f1-e512b5d33ae6-kube-api-access-vsbhg" (OuterVolumeSpecName: "kube-api-access-vsbhg") pod "66691427-c48c-4dbe-b5f1-e512b5d33ae6" (UID: "66691427-c48c-4dbe-b5f1-e512b5d33ae6"). InnerVolumeSpecName "kube-api-access-vsbhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.875700 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.875945 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a07b6c09-179a-4562-ac56-5ac471127525-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.876010 4744 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.876130 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/66691427-c48c-4dbe-b5f1-e512b5d33ae6-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.876216 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gpp2\" (UniqueName: \"kubernetes.io/projected/a07b6c09-179a-4562-ac56-5ac471127525-kube-api-access-2gpp2\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.876280 4744 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-client-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.876339 4744 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a07b6c09-179a-4562-ac56-5ac471127525-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.876398 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsbhg\" (UniqueName: \"kubernetes.io/projected/66691427-c48c-4dbe-b5f1-e512b5d33ae6-kube-api-access-vsbhg\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:03 crc kubenswrapper[4744]: I0106 14:41:03.876454 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66691427-c48c-4dbe-b5f1-e512b5d33ae6-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.295397 4744 generic.go:334] "Generic (PLEG): container finished" podID="a07b6c09-179a-4562-ac56-5ac471127525" containerID="3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9" exitCode=0 Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.295694 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.299346 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" event={"ID":"a07b6c09-179a-4562-ac56-5ac471127525","Type":"ContainerDied","Data":"3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9"} Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.299429 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx" event={"ID":"a07b6c09-179a-4562-ac56-5ac471127525","Type":"ContainerDied","Data":"4a308e65c73e4c9ae675b1de75d576b34239885dd3a1a6138e1d34b237fd91e3"} Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.299458 4744 scope.go:117] "RemoveContainer" containerID="3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9" Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.303924 4744 generic.go:334] "Generic (PLEG): container finished" podID="66691427-c48c-4dbe-b5f1-e512b5d33ae6" containerID="5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368" exitCode=0 Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.303969 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" event={"ID":"66691427-c48c-4dbe-b5f1-e512b5d33ae6","Type":"ContainerDied","Data":"5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368"} Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.303998 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" event={"ID":"66691427-c48c-4dbe-b5f1-e512b5d33ae6","Type":"ContainerDied","Data":"ae46b6e36bcc877a17a78f96d9934f7691d5276b7a18c42069fad30bf6725f35"} Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.304064 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f55d8b9f-wsn84" Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.342789 4744 scope.go:117] "RemoveContainer" containerID="3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9" Jan 06 14:41:04 crc kubenswrapper[4744]: E0106 14:41:04.345459 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9\": container with ID starting with 3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9 not found: ID does not exist" containerID="3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9" Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.345545 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9"} err="failed to get container status \"3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9\": rpc error: code = NotFound desc = could not find container \"3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9\": container with ID starting with 3614b1f8f3fc26e381422342a689e4ef4a88c1df6ca757eda8e04392f73691c9 not found: ID does not exist" Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.345602 4744 scope.go:117] "RemoveContainer" containerID="5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368" Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.351503 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66f55d8b9f-wsn84"] Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.359044 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-66f55d8b9f-wsn84"] Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.363041 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx"] Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.366637 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd6f6d45-7kcmx"] Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.371083 4744 scope.go:117] "RemoveContainer" containerID="5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368" Jan 06 14:41:04 crc kubenswrapper[4744]: E0106 14:41:04.371596 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368\": container with ID starting with 5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368 not found: ID does not exist" containerID="5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368" Jan 06 14:41:04 crc kubenswrapper[4744]: I0106 14:41:04.371630 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368"} err="failed to get container status \"5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368\": rpc error: code = NotFound desc = could not find container \"5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368\": container with ID starting with 5fbd5ae60fc2d024fa13612d19f422d020fff6848aebd846eeb13c48a624e368 not found: ID does not exist" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.005660 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-688c459784-g8whl"] Jan 06 14:41:05 crc kubenswrapper[4744]: E0106 14:41:05.006098 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a07b6c09-179a-4562-ac56-5ac471127525" containerName="route-controller-manager" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.006130 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a07b6c09-179a-4562-ac56-5ac471127525" containerName="route-controller-manager" Jan 06 14:41:05 crc kubenswrapper[4744]: E0106 14:41:05.006184 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66691427-c48c-4dbe-b5f1-e512b5d33ae6" containerName="controller-manager" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.006198 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="66691427-c48c-4dbe-b5f1-e512b5d33ae6" containerName="controller-manager" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.006398 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a07b6c09-179a-4562-ac56-5ac471127525" containerName="route-controller-manager" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.006425 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="66691427-c48c-4dbe-b5f1-e512b5d33ae6" containerName="controller-manager" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.007122 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.008938 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.009682 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.015932 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.015943 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.017360 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.017789 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.018059 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.018062 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.018138 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.018235 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.018257 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.018651 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.019987 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.020004 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.026087 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.029019 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-688c459784-g8whl"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.031895 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.091239 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-serving-cert\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.091293 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-client-ca\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.091330 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-serving-cert\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.091353 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brmv8\" (UniqueName: \"kubernetes.io/projected/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-kube-api-access-brmv8\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.091487 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msk9v\" (UniqueName: \"kubernetes.io/projected/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-kube-api-access-msk9v\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.091666 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-proxy-ca-bundles\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.091789 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-config\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.091939 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-config\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.091988 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-client-ca\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.175348 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c8458"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.175889 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c8458" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" containerName="registry-server" containerID="cri-o://9f828e570440772fe4175a793636c299683646545d0de22a97fdd931f3ac0a14" gracePeriod=30 Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.184283 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4cwvb"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.184617 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4cwvb" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerName="registry-server" containerID="cri-o://f79f3e3e7070ccaaf2c98dba729e886d3e24853359fa99139a27d0a23a1d41ac" gracePeriod=30 Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.193364 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgpw"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.193843 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" podUID="f0690b11-315a-4639-8455-d3323eb7e042" containerName="marketplace-operator" containerID="cri-o://a7c178b0865fc44d01d8d02f6a72c64a84c203e52857b42d4d324229547e5614" gracePeriod=30 Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.197620 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-client-ca\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.197703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-serving-cert\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.197743 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-client-ca\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.197797 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-serving-cert\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.197848 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brmv8\" (UniqueName: \"kubernetes.io/projected/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-kube-api-access-brmv8\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.197916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msk9v\" (UniqueName: \"kubernetes.io/projected/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-kube-api-access-msk9v\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.197953 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-proxy-ca-bundles\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.198037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-config\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.198125 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-config\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.198228 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hxcm"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.198572 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4hxcm" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerName="registry-server" containerID="cri-o://e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380" gracePeriod=30 Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.199300 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-client-ca\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.199681 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-client-ca\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.200002 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-config\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.200024 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-proxy-ca-bundles\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.205927 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-config\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.210245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-serving-cert\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.215836 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bh4q"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.216267 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4bh4q" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerName="registry-server" containerID="cri-o://64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1" gracePeriod=30 Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.221558 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-serving-cert\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.221759 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wsc5g"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.224287 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.230627 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brmv8\" (UniqueName: \"kubernetes.io/projected/53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d-kube-api-access-brmv8\") pod \"controller-manager-688c459784-g8whl\" (UID: \"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d\") " pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.235204 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msk9v\" (UniqueName: \"kubernetes.io/projected/8ed4bc19-0cbb-41fe-9956-5aef0d4c287b-kube-api-access-msk9v\") pod \"route-controller-manager-64bb695654-t9m84\" (UID: \"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b\") " pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.256728 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wsc5g"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.300373 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6c65897-076d-4e90-a28a-2989a57e11be-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wsc5g\" (UID: \"e6c65897-076d-4e90-a28a-2989a57e11be\") " pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.300479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9qqh\" (UniqueName: \"kubernetes.io/projected/e6c65897-076d-4e90-a28a-2989a57e11be-kube-api-access-c9qqh\") pod \"marketplace-operator-79b997595-wsc5g\" (UID: \"e6c65897-076d-4e90-a28a-2989a57e11be\") " pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.300546 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e6c65897-076d-4e90-a28a-2989a57e11be-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wsc5g\" (UID: \"e6c65897-076d-4e90-a28a-2989a57e11be\") " pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.314563 4744 generic.go:334] "Generic (PLEG): container finished" podID="f0690b11-315a-4639-8455-d3323eb7e042" containerID="a7c178b0865fc44d01d8d02f6a72c64a84c203e52857b42d4d324229547e5614" exitCode=0 Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.314623 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" event={"ID":"f0690b11-315a-4639-8455-d3323eb7e042","Type":"ContainerDied","Data":"a7c178b0865fc44d01d8d02f6a72c64a84c203e52857b42d4d324229547e5614"} Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.322798 4744 generic.go:334] "Generic (PLEG): container finished" podID="ead06eb7-e94b-400a-b568-1871fea3e807" containerID="9f828e570440772fe4175a793636c299683646545d0de22a97fdd931f3ac0a14" exitCode=0 Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.323096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8458" event={"ID":"ead06eb7-e94b-400a-b568-1871fea3e807","Type":"ContainerDied","Data":"9f828e570440772fe4175a793636c299683646545d0de22a97fdd931f3ac0a14"} Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.324963 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.331722 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.347882 4744 generic.go:334] "Generic (PLEG): container finished" podID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerID="f79f3e3e7070ccaaf2c98dba729e886d3e24853359fa99139a27d0a23a1d41ac" exitCode=0 Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.347931 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cwvb" event={"ID":"14d7cfd7-181f-4d79-8951-27d300d49a8f","Type":"ContainerDied","Data":"f79f3e3e7070ccaaf2c98dba729e886d3e24853359fa99139a27d0a23a1d41ac"} Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.401679 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6c65897-076d-4e90-a28a-2989a57e11be-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wsc5g\" (UID: \"e6c65897-076d-4e90-a28a-2989a57e11be\") " pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.401744 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9qqh\" (UniqueName: \"kubernetes.io/projected/e6c65897-076d-4e90-a28a-2989a57e11be-kube-api-access-c9qqh\") pod \"marketplace-operator-79b997595-wsc5g\" (UID: \"e6c65897-076d-4e90-a28a-2989a57e11be\") " pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.401771 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e6c65897-076d-4e90-a28a-2989a57e11be-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wsc5g\" (UID: \"e6c65897-076d-4e90-a28a-2989a57e11be\") " pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.403005 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e6c65897-076d-4e90-a28a-2989a57e11be-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wsc5g\" (UID: \"e6c65897-076d-4e90-a28a-2989a57e11be\") " pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.404780 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/e6c65897-076d-4e90-a28a-2989a57e11be-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wsc5g\" (UID: \"e6c65897-076d-4e90-a28a-2989a57e11be\") " pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.419326 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9qqh\" (UniqueName: \"kubernetes.io/projected/e6c65897-076d-4e90-a28a-2989a57e11be-kube-api-access-c9qqh\") pod \"marketplace-operator-79b997595-wsc5g\" (UID: \"e6c65897-076d-4e90-a28a-2989a57e11be\") " pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.607463 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.634759 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.646469 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.686648 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.691824 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.707737 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-operator-metrics\") pod \"f0690b11-315a-4639-8455-d3323eb7e042\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.707800 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jthl\" (UniqueName: \"kubernetes.io/projected/ead06eb7-e94b-400a-b568-1871fea3e807-kube-api-access-2jthl\") pod \"ead06eb7-e94b-400a-b568-1871fea3e807\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.707837 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-catalog-content\") pod \"ead06eb7-e94b-400a-b568-1871fea3e807\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.707918 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-utilities\") pod \"ead06eb7-e94b-400a-b568-1871fea3e807\" (UID: \"ead06eb7-e94b-400a-b568-1871fea3e807\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.707938 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xldjt\" (UniqueName: \"kubernetes.io/projected/f0690b11-315a-4639-8455-d3323eb7e042-kube-api-access-xldjt\") pod \"f0690b11-315a-4639-8455-d3323eb7e042\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.707988 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-trusted-ca\") pod \"f0690b11-315a-4639-8455-d3323eb7e042\" (UID: \"f0690b11-315a-4639-8455-d3323eb7e042\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.708869 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-utilities" (OuterVolumeSpecName: "utilities") pod "ead06eb7-e94b-400a-b568-1871fea3e807" (UID: "ead06eb7-e94b-400a-b568-1871fea3e807"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.709198 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f0690b11-315a-4639-8455-d3323eb7e042" (UID: "f0690b11-315a-4639-8455-d3323eb7e042"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.709979 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.711493 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0690b11-315a-4639-8455-d3323eb7e042-kube-api-access-xldjt" (OuterVolumeSpecName: "kube-api-access-xldjt") pod "f0690b11-315a-4639-8455-d3323eb7e042" (UID: "f0690b11-315a-4639-8455-d3323eb7e042"). InnerVolumeSpecName "kube-api-access-xldjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.712595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ead06eb7-e94b-400a-b568-1871fea3e807-kube-api-access-2jthl" (OuterVolumeSpecName: "kube-api-access-2jthl") pod "ead06eb7-e94b-400a-b568-1871fea3e807" (UID: "ead06eb7-e94b-400a-b568-1871fea3e807"). InnerVolumeSpecName "kube-api-access-2jthl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.713513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f0690b11-315a-4639-8455-d3323eb7e042" (UID: "f0690b11-315a-4639-8455-d3323eb7e042"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.720285 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66691427-c48c-4dbe-b5f1-e512b5d33ae6" path="/var/lib/kubelet/pods/66691427-c48c-4dbe-b5f1-e512b5d33ae6/volumes" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.720939 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a07b6c09-179a-4562-ac56-5ac471127525" path="/var/lib/kubelet/pods/a07b6c09-179a-4562-ac56-5ac471127525/volumes" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.792629 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ead06eb7-e94b-400a-b568-1871fea3e807" (UID: "ead06eb7-e94b-400a-b568-1871fea3e807"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.808888 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-catalog-content\") pod \"6fe6e90c-e807-441c-b130-15a7368b87f0\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.808940 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-catalog-content\") pod \"14d7cfd7-181f-4d79-8951-27d300d49a8f\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809001 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-catalog-content\") pod \"408386a1-c1c3-4b95-8bce-4c60433159a0\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809029 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-utilities\") pod \"14d7cfd7-181f-4d79-8951-27d300d49a8f\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809076 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68bnt\" (UniqueName: \"kubernetes.io/projected/408386a1-c1c3-4b95-8bce-4c60433159a0-kube-api-access-68bnt\") pod \"408386a1-c1c3-4b95-8bce-4c60433159a0\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809093 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-utilities\") pod \"408386a1-c1c3-4b95-8bce-4c60433159a0\" (UID: \"408386a1-c1c3-4b95-8bce-4c60433159a0\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809128 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98h5j\" (UniqueName: \"kubernetes.io/projected/14d7cfd7-181f-4d79-8951-27d300d49a8f-kube-api-access-98h5j\") pod \"14d7cfd7-181f-4d79-8951-27d300d49a8f\" (UID: \"14d7cfd7-181f-4d79-8951-27d300d49a8f\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809145 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-utilities\") pod \"6fe6e90c-e807-441c-b130-15a7368b87f0\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809173 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jh2lp\" (UniqueName: \"kubernetes.io/projected/6fe6e90c-e807-441c-b130-15a7368b87f0-kube-api-access-jh2lp\") pod \"6fe6e90c-e807-441c-b130-15a7368b87f0\" (UID: \"6fe6e90c-e807-441c-b130-15a7368b87f0\") " Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809418 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809430 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xldjt\" (UniqueName: \"kubernetes.io/projected/f0690b11-315a-4639-8455-d3323eb7e042-kube-api-access-xldjt\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809440 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809451 4744 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f0690b11-315a-4639-8455-d3323eb7e042-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809461 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jthl\" (UniqueName: \"kubernetes.io/projected/ead06eb7-e94b-400a-b568-1871fea3e807-kube-api-access-2jthl\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.809469 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead06eb7-e94b-400a-b568-1871fea3e807-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.814870 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-utilities" (OuterVolumeSpecName: "utilities") pod "408386a1-c1c3-4b95-8bce-4c60433159a0" (UID: "408386a1-c1c3-4b95-8bce-4c60433159a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.815006 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-utilities" (OuterVolumeSpecName: "utilities") pod "6fe6e90c-e807-441c-b130-15a7368b87f0" (UID: "6fe6e90c-e807-441c-b130-15a7368b87f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.815080 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-utilities" (OuterVolumeSpecName: "utilities") pod "14d7cfd7-181f-4d79-8951-27d300d49a8f" (UID: "14d7cfd7-181f-4d79-8951-27d300d49a8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.816871 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fe6e90c-e807-441c-b130-15a7368b87f0-kube-api-access-jh2lp" (OuterVolumeSpecName: "kube-api-access-jh2lp") pod "6fe6e90c-e807-441c-b130-15a7368b87f0" (UID: "6fe6e90c-e807-441c-b130-15a7368b87f0"). InnerVolumeSpecName "kube-api-access-jh2lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.817397 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/408386a1-c1c3-4b95-8bce-4c60433159a0-kube-api-access-68bnt" (OuterVolumeSpecName: "kube-api-access-68bnt") pod "408386a1-c1c3-4b95-8bce-4c60433159a0" (UID: "408386a1-c1c3-4b95-8bce-4c60433159a0"). InnerVolumeSpecName "kube-api-access-68bnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.817442 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14d7cfd7-181f-4d79-8951-27d300d49a8f-kube-api-access-98h5j" (OuterVolumeSpecName: "kube-api-access-98h5j") pod "14d7cfd7-181f-4d79-8951-27d300d49a8f" (UID: "14d7cfd7-181f-4d79-8951-27d300d49a8f"). InnerVolumeSpecName "kube-api-access-98h5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.844103 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fe6e90c-e807-441c-b130-15a7368b87f0" (UID: "6fe6e90c-e807-441c-b130-15a7368b87f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.857420 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wsc5g"] Jan 06 14:41:05 crc kubenswrapper[4744]: W0106 14:41:05.859050 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6c65897_076d_4e90_a28a_2989a57e11be.slice/crio-20eb3e45fa5d1c504300bdf71242b58aaa7f4573a72426993c7849f72703ad6f WatchSource:0}: Error finding container 20eb3e45fa5d1c504300bdf71242b58aaa7f4573a72426993c7849f72703ad6f: Status 404 returned error can't find the container with id 20eb3e45fa5d1c504300bdf71242b58aaa7f4573a72426993c7849f72703ad6f Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.880958 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14d7cfd7-181f-4d79-8951-27d300d49a8f" (UID: "14d7cfd7-181f-4d79-8951-27d300d49a8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.910788 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.910843 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.910859 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68bnt\" (UniqueName: \"kubernetes.io/projected/408386a1-c1c3-4b95-8bce-4c60433159a0-kube-api-access-68bnt\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.910873 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98h5j\" (UniqueName: \"kubernetes.io/projected/14d7cfd7-181f-4d79-8951-27d300d49a8f-kube-api-access-98h5j\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.910886 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.910902 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jh2lp\" (UniqueName: \"kubernetes.io/projected/6fe6e90c-e807-441c-b130-15a7368b87f0-kube-api-access-jh2lp\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.910915 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fe6e90c-e807-441c-b130-15a7368b87f0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.910927 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14d7cfd7-181f-4d79-8951-27d300d49a8f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.912390 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.930229 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-688c459784-g8whl"] Jan 06 14:41:05 crc kubenswrapper[4744]: I0106 14:41:05.970585 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "408386a1-c1c3-4b95-8bce-4c60433159a0" (UID: "408386a1-c1c3-4b95-8bce-4c60433159a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.013133 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/408386a1-c1c3-4b95-8bce-4c60433159a0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.355386 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c8458" event={"ID":"ead06eb7-e94b-400a-b568-1871fea3e807","Type":"ContainerDied","Data":"eba2ea840fc97fd2f00f616943413ccf0f41f571ce13d91dd17a2cd6d3794639"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.355469 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c8458" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.355503 4744 scope.go:117] "RemoveContainer" containerID="9f828e570440772fe4175a793636c299683646545d0de22a97fdd931f3ac0a14" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.362719 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4cwvb" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.362852 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4cwvb" event={"ID":"14d7cfd7-181f-4d79-8951-27d300d49a8f","Type":"ContainerDied","Data":"0fa264b620a099372f580860981c19bc14169722a645de9882d9f5f7322972a9"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.372201 4744 generic.go:334] "Generic (PLEG): container finished" podID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerID="e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380" exitCode=0 Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.372281 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hxcm" event={"ID":"6fe6e90c-e807-441c-b130-15a7368b87f0","Type":"ContainerDied","Data":"e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.372321 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hxcm" event={"ID":"6fe6e90c-e807-441c-b130-15a7368b87f0","Type":"ContainerDied","Data":"d5f78c308bd69589e5535d1f5f1b82f57e3784e9238c57f04e821fd57f94e787"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.372436 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hxcm" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.378776 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" event={"ID":"e6c65897-076d-4e90-a28a-2989a57e11be","Type":"ContainerStarted","Data":"f262dcbf1fcafe16f9964b2742ddacd7b0965d137fabd7dc08927541fcfb8210"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.378832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" event={"ID":"e6c65897-076d-4e90-a28a-2989a57e11be","Type":"ContainerStarted","Data":"20eb3e45fa5d1c504300bdf71242b58aaa7f4573a72426993c7849f72703ad6f"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.379095 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.381184 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" event={"ID":"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b","Type":"ContainerStarted","Data":"3a34046ba1eb14c0f707b198ac3074bf62e87564d303bb0c44329ad8f7ea44e2"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.381234 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" event={"ID":"8ed4bc19-0cbb-41fe-9956-5aef0d4c287b","Type":"ContainerStarted","Data":"9794dbfedfebf2c51f7cdf1c8cb845cb64eec2a6eb1b9ea85db3d3508f5c25e2"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.381855 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.385046 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.386286 4744 scope.go:117] "RemoveContainer" containerID="d4677c4cb280809b533e20341865dcd6a84861cade2772820f6882c5a3665550" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.394339 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-688c459784-g8whl" event={"ID":"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d","Type":"ContainerStarted","Data":"599280e9be00dcb7171d9453c4be6b9edba737287031645fe70d825316348245"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.394384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-688c459784-g8whl" event={"ID":"53c48f0c-ce19-4ba6-9f51-2be2f1a10b9d","Type":"ContainerStarted","Data":"61154384f5a86084ab3ea2005b98a3231d6b5b502bbf01694b535909fd42fd42"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.394729 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.396951 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" event={"ID":"f0690b11-315a-4639-8455-d3323eb7e042","Type":"ContainerDied","Data":"7774f1f0e040ab54031de788f52146dd8c7625f1210aeac6284127fce585cdc5"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.397008 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kzgpw" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.414228 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wsc5g" podStartSLOduration=1.414201059 podStartE2EDuration="1.414201059s" podCreationTimestamp="2026-01-06 14:41:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:41:06.411445315 +0000 UTC m=+263.038911643" watchObservedRunningTime="2026-01-06 14:41:06.414201059 +0000 UTC m=+263.041667377" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.420492 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-688c459784-g8whl" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.420982 4744 generic.go:334] "Generic (PLEG): container finished" podID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerID="64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1" exitCode=0 Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.421031 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bh4q" event={"ID":"408386a1-c1c3-4b95-8bce-4c60433159a0","Type":"ContainerDied","Data":"64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.421065 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bh4q" event={"ID":"408386a1-c1c3-4b95-8bce-4c60433159a0","Type":"ContainerDied","Data":"72a8d03499b533e3150127bb2e0e21cbf159a8a4d79e5760e326a06ed473a67c"} Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.421154 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bh4q" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.428473 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hxcm"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.432723 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hxcm"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.436337 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.437906 4744 scope.go:117] "RemoveContainer" containerID="9d6778973ec5c041f7fa6f29df88279be671a5ecd0f5bee8f4ed8f9697996e93" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.444727 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" podStartSLOduration=3.44471531 podStartE2EDuration="3.44471531s" podCreationTimestamp="2026-01-06 14:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:41:06.441458983 +0000 UTC m=+263.068925301" watchObservedRunningTime="2026-01-06 14:41:06.44471531 +0000 UTC m=+263.072181628" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.453741 4744 scope.go:117] "RemoveContainer" containerID="f79f3e3e7070ccaaf2c98dba729e886d3e24853359fa99139a27d0a23a1d41ac" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.477334 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-688c459784-g8whl" podStartSLOduration=3.477307078 podStartE2EDuration="3.477307078s" podCreationTimestamp="2026-01-06 14:41:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:41:06.46551509 +0000 UTC m=+263.092981408" watchObservedRunningTime="2026-01-06 14:41:06.477307078 +0000 UTC m=+263.104773396" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.481266 4744 scope.go:117] "RemoveContainer" containerID="2ecf8435a9157687bd7b0e63b5a0528dcef2691ab79a21eeca329addb84772c0" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.492264 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4cwvb"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.494036 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4cwvb"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.502336 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c8458"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.506727 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c8458"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.508828 4744 scope.go:117] "RemoveContainer" containerID="c4209be2450e3939a345ed77b5b447158f39954aa943c4d1e0bd23ca4a8d1c3a" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.517287 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgpw"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.528975 4744 scope.go:117] "RemoveContainer" containerID="e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.535272 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kzgpw"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.546350 4744 scope.go:117] "RemoveContainer" containerID="56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.560221 4744 scope.go:117] "RemoveContainer" containerID="56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.593726 4744 scope.go:117] "RemoveContainer" containerID="e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380" Jan 06 14:41:06 crc kubenswrapper[4744]: E0106 14:41:06.594351 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380\": container with ID starting with e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380 not found: ID does not exist" containerID="e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.594397 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380"} err="failed to get container status \"e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380\": rpc error: code = NotFound desc = could not find container \"e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380\": container with ID starting with e8d7f453a1222c6699684457c59d332e27e27191a8449d1fbfa34fab077bc380 not found: ID does not exist" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.594425 4744 scope.go:117] "RemoveContainer" containerID="56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176" Jan 06 14:41:06 crc kubenswrapper[4744]: E0106 14:41:06.594667 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176\": container with ID starting with 56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176 not found: ID does not exist" containerID="56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.594689 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176"} err="failed to get container status \"56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176\": rpc error: code = NotFound desc = could not find container \"56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176\": container with ID starting with 56bdacbdaa96aa3a99e12eecd96508b7d6566cc789781c168af43ed7ef357176 not found: ID does not exist" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.594703 4744 scope.go:117] "RemoveContainer" containerID="56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb" Jan 06 14:41:06 crc kubenswrapper[4744]: E0106 14:41:06.595844 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb\": container with ID starting with 56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb not found: ID does not exist" containerID="56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.595872 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb"} err="failed to get container status \"56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb\": rpc error: code = NotFound desc = could not find container \"56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb\": container with ID starting with 56c45ddffb4ffa64661abcfee370bf54cfeef51b470a860004f10197b5e8cabb not found: ID does not exist" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.595888 4744 scope.go:117] "RemoveContainer" containerID="a7c178b0865fc44d01d8d02f6a72c64a84c203e52857b42d4d324229547e5614" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.597220 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bh4q"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.618101 4744 scope.go:117] "RemoveContainer" containerID="64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.623709 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4bh4q"] Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.648242 4744 scope.go:117] "RemoveContainer" containerID="66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.673205 4744 scope.go:117] "RemoveContainer" containerID="afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.689350 4744 scope.go:117] "RemoveContainer" containerID="64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1" Jan 06 14:41:06 crc kubenswrapper[4744]: E0106 14:41:06.689728 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1\": container with ID starting with 64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1 not found: ID does not exist" containerID="64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.689769 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1"} err="failed to get container status \"64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1\": rpc error: code = NotFound desc = could not find container \"64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1\": container with ID starting with 64c713e3cd0e7b69f2a2a70cebb83a5e046135a467c5f89ff098932e7723c2a1 not found: ID does not exist" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.689794 4744 scope.go:117] "RemoveContainer" containerID="66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5" Jan 06 14:41:06 crc kubenswrapper[4744]: E0106 14:41:06.694804 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5\": container with ID starting with 66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5 not found: ID does not exist" containerID="66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.694839 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5"} err="failed to get container status \"66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5\": rpc error: code = NotFound desc = could not find container \"66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5\": container with ID starting with 66dbcd9c995704b147a7142da9a42134c592ac9ba39ab66ffdb1893dd62271b5 not found: ID does not exist" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.694861 4744 scope.go:117] "RemoveContainer" containerID="afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30" Jan 06 14:41:06 crc kubenswrapper[4744]: E0106 14:41:06.695295 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30\": container with ID starting with afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30 not found: ID does not exist" containerID="afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30" Jan 06 14:41:06 crc kubenswrapper[4744]: I0106 14:41:06.695322 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30"} err="failed to get container status \"afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30\": rpc error: code = NotFound desc = could not find container \"afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30\": container with ID starting with afa3415caa6dbb61ecb819648fdac573d32cb6727f61bea86b445d5d95008f30 not found: ID does not exist" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401404 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pssbh"] Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401676 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerName="extract-utilities" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401693 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerName="extract-utilities" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401708 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401718 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401728 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerName="extract-content" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401737 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerName="extract-content" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401747 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerName="extract-content" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401756 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerName="extract-content" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401768 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" containerName="extract-utilities" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401777 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" containerName="extract-utilities" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401787 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerName="extract-content" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401796 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerName="extract-content" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401809 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401819 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401832 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401842 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401857 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerName="extract-utilities" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401867 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerName="extract-utilities" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401877 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" containerName="extract-content" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401887 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" containerName="extract-content" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401903 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0690b11-315a-4639-8455-d3323eb7e042" containerName="marketplace-operator" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401913 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0690b11-315a-4639-8455-d3323eb7e042" containerName="marketplace-operator" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401930 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401940 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: E0106 14:41:07.401978 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerName="extract-utilities" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.401990 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerName="extract-utilities" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.402111 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.402124 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0690b11-315a-4639-8455-d3323eb7e042" containerName="marketplace-operator" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.402136 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.402151 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.402186 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" containerName="registry-server" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.403154 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.405901 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.419406 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pssbh"] Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.531099 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mcxr\" (UniqueName: \"kubernetes.io/projected/d470584a-0f48-436b-9ead-06f8947b7115-kube-api-access-7mcxr\") pod \"redhat-marketplace-pssbh\" (UID: \"d470584a-0f48-436b-9ead-06f8947b7115\") " pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.531203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d470584a-0f48-436b-9ead-06f8947b7115-catalog-content\") pod \"redhat-marketplace-pssbh\" (UID: \"d470584a-0f48-436b-9ead-06f8947b7115\") " pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.531239 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d470584a-0f48-436b-9ead-06f8947b7115-utilities\") pod \"redhat-marketplace-pssbh\" (UID: \"d470584a-0f48-436b-9ead-06f8947b7115\") " pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.592517 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vrxvs"] Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.594395 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.597843 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.610482 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vrxvs"] Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.631980 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d470584a-0f48-436b-9ead-06f8947b7115-utilities\") pod \"redhat-marketplace-pssbh\" (UID: \"d470584a-0f48-436b-9ead-06f8947b7115\") " pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.632060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mcxr\" (UniqueName: \"kubernetes.io/projected/d470584a-0f48-436b-9ead-06f8947b7115-kube-api-access-7mcxr\") pod \"redhat-marketplace-pssbh\" (UID: \"d470584a-0f48-436b-9ead-06f8947b7115\") " pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.632106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d470584a-0f48-436b-9ead-06f8947b7115-catalog-content\") pod \"redhat-marketplace-pssbh\" (UID: \"d470584a-0f48-436b-9ead-06f8947b7115\") " pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.633267 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d470584a-0f48-436b-9ead-06f8947b7115-catalog-content\") pod \"redhat-marketplace-pssbh\" (UID: \"d470584a-0f48-436b-9ead-06f8947b7115\") " pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.633523 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d470584a-0f48-436b-9ead-06f8947b7115-utilities\") pod \"redhat-marketplace-pssbh\" (UID: \"d470584a-0f48-436b-9ead-06f8947b7115\") " pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.652368 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mcxr\" (UniqueName: \"kubernetes.io/projected/d470584a-0f48-436b-9ead-06f8947b7115-kube-api-access-7mcxr\") pod \"redhat-marketplace-pssbh\" (UID: \"d470584a-0f48-436b-9ead-06f8947b7115\") " pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.719394 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14d7cfd7-181f-4d79-8951-27d300d49a8f" path="/var/lib/kubelet/pods/14d7cfd7-181f-4d79-8951-27d300d49a8f/volumes" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.721049 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="408386a1-c1c3-4b95-8bce-4c60433159a0" path="/var/lib/kubelet/pods/408386a1-c1c3-4b95-8bce-4c60433159a0/volumes" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.722364 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fe6e90c-e807-441c-b130-15a7368b87f0" path="/var/lib/kubelet/pods/6fe6e90c-e807-441c-b130-15a7368b87f0/volumes" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.725066 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.725236 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ead06eb7-e94b-400a-b568-1871fea3e807" path="/var/lib/kubelet/pods/ead06eb7-e94b-400a-b568-1871fea3e807/volumes" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.727473 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0690b11-315a-4639-8455-d3323eb7e042" path="/var/lib/kubelet/pods/f0690b11-315a-4639-8455-d3323eb7e042/volumes" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.733769 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5rd2\" (UniqueName: \"kubernetes.io/projected/78579813-4567-46c4-8353-4e6418ba7722-kube-api-access-z5rd2\") pod \"redhat-operators-vrxvs\" (UID: \"78579813-4567-46c4-8353-4e6418ba7722\") " pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.733809 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78579813-4567-46c4-8353-4e6418ba7722-utilities\") pod \"redhat-operators-vrxvs\" (UID: \"78579813-4567-46c4-8353-4e6418ba7722\") " pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.733830 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78579813-4567-46c4-8353-4e6418ba7722-catalog-content\") pod \"redhat-operators-vrxvs\" (UID: \"78579813-4567-46c4-8353-4e6418ba7722\") " pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.835349 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5rd2\" (UniqueName: \"kubernetes.io/projected/78579813-4567-46c4-8353-4e6418ba7722-kube-api-access-z5rd2\") pod \"redhat-operators-vrxvs\" (UID: \"78579813-4567-46c4-8353-4e6418ba7722\") " pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.835412 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78579813-4567-46c4-8353-4e6418ba7722-utilities\") pod \"redhat-operators-vrxvs\" (UID: \"78579813-4567-46c4-8353-4e6418ba7722\") " pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.835444 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78579813-4567-46c4-8353-4e6418ba7722-catalog-content\") pod \"redhat-operators-vrxvs\" (UID: \"78579813-4567-46c4-8353-4e6418ba7722\") " pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.836467 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78579813-4567-46c4-8353-4e6418ba7722-catalog-content\") pod \"redhat-operators-vrxvs\" (UID: \"78579813-4567-46c4-8353-4e6418ba7722\") " pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.838306 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78579813-4567-46c4-8353-4e6418ba7722-utilities\") pod \"redhat-operators-vrxvs\" (UID: \"78579813-4567-46c4-8353-4e6418ba7722\") " pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.863800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5rd2\" (UniqueName: \"kubernetes.io/projected/78579813-4567-46c4-8353-4e6418ba7722-kube-api-access-z5rd2\") pod \"redhat-operators-vrxvs\" (UID: \"78579813-4567-46c4-8353-4e6418ba7722\") " pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.920342 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:07 crc kubenswrapper[4744]: I0106 14:41:07.968884 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pssbh"] Jan 06 14:41:07 crc kubenswrapper[4744]: W0106 14:41:07.979090 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd470584a_0f48_436b_9ead_06f8947b7115.slice/crio-49da5efeb13021b45f6dd7783f8da78877736d74ae616af0739720d165c7bf23 WatchSource:0}: Error finding container 49da5efeb13021b45f6dd7783f8da78877736d74ae616af0739720d165c7bf23: Status 404 returned error can't find the container with id 49da5efeb13021b45f6dd7783f8da78877736d74ae616af0739720d165c7bf23 Jan 06 14:41:08 crc kubenswrapper[4744]: I0106 14:41:08.110647 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vrxvs"] Jan 06 14:41:08 crc kubenswrapper[4744]: I0106 14:41:08.475895 4744 generic.go:334] "Generic (PLEG): container finished" podID="78579813-4567-46c4-8353-4e6418ba7722" containerID="e9b09c439e035d1d5c86107c41de889af4b033d5f3fa3a0ce0e2d7d3fc4de3a2" exitCode=0 Jan 06 14:41:08 crc kubenswrapper[4744]: I0106 14:41:08.476481 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrxvs" event={"ID":"78579813-4567-46c4-8353-4e6418ba7722","Type":"ContainerDied","Data":"e9b09c439e035d1d5c86107c41de889af4b033d5f3fa3a0ce0e2d7d3fc4de3a2"} Jan 06 14:41:08 crc kubenswrapper[4744]: I0106 14:41:08.476559 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrxvs" event={"ID":"78579813-4567-46c4-8353-4e6418ba7722","Type":"ContainerStarted","Data":"dd137d41f7e22407b1d616faf052df3696af9a52c3fd43e27f69f206825c2687"} Jan 06 14:41:08 crc kubenswrapper[4744]: I0106 14:41:08.479377 4744 generic.go:334] "Generic (PLEG): container finished" podID="d470584a-0f48-436b-9ead-06f8947b7115" containerID="8885c8a2efd1bd5f576fa4ef6f0f2003dc3144e9ea5654d9a38da3985fa5bb46" exitCode=0 Jan 06 14:41:08 crc kubenswrapper[4744]: I0106 14:41:08.480367 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pssbh" event={"ID":"d470584a-0f48-436b-9ead-06f8947b7115","Type":"ContainerDied","Data":"8885c8a2efd1bd5f576fa4ef6f0f2003dc3144e9ea5654d9a38da3985fa5bb46"} Jan 06 14:41:08 crc kubenswrapper[4744]: I0106 14:41:08.480417 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pssbh" event={"ID":"d470584a-0f48-436b-9ead-06f8947b7115","Type":"ContainerStarted","Data":"49da5efeb13021b45f6dd7783f8da78877736d74ae616af0739720d165c7bf23"} Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.486667 4744 generic.go:334] "Generic (PLEG): container finished" podID="d470584a-0f48-436b-9ead-06f8947b7115" containerID="cb8bbff3824fb63e510f32b85bb41275477a761888041ff13b2570571863d8c4" exitCode=0 Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.486745 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pssbh" event={"ID":"d470584a-0f48-436b-9ead-06f8947b7115","Type":"ContainerDied","Data":"cb8bbff3824fb63e510f32b85bb41275477a761888041ff13b2570571863d8c4"} Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.491314 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrxvs" event={"ID":"78579813-4567-46c4-8353-4e6418ba7722","Type":"ContainerStarted","Data":"d9839f2c061ad2661413cd17f45ae407deb018f452ed883f8d508e9038720e75"} Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.794131 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8755w"] Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.795995 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.799288 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.812082 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8755w"] Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.861845 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/889a8fc3-2966-446c-96d1-fa9203b19280-utilities\") pod \"community-operators-8755w\" (UID: \"889a8fc3-2966-446c-96d1-fa9203b19280\") " pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.861991 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6k9f\" (UniqueName: \"kubernetes.io/projected/889a8fc3-2966-446c-96d1-fa9203b19280-kube-api-access-p6k9f\") pod \"community-operators-8755w\" (UID: \"889a8fc3-2966-446c-96d1-fa9203b19280\") " pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.862320 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/889a8fc3-2966-446c-96d1-fa9203b19280-catalog-content\") pod \"community-operators-8755w\" (UID: \"889a8fc3-2966-446c-96d1-fa9203b19280\") " pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.964492 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/889a8fc3-2966-446c-96d1-fa9203b19280-catalog-content\") pod \"community-operators-8755w\" (UID: \"889a8fc3-2966-446c-96d1-fa9203b19280\") " pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.964593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/889a8fc3-2966-446c-96d1-fa9203b19280-utilities\") pod \"community-operators-8755w\" (UID: \"889a8fc3-2966-446c-96d1-fa9203b19280\") " pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.964642 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6k9f\" (UniqueName: \"kubernetes.io/projected/889a8fc3-2966-446c-96d1-fa9203b19280-kube-api-access-p6k9f\") pod \"community-operators-8755w\" (UID: \"889a8fc3-2966-446c-96d1-fa9203b19280\") " pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.964998 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/889a8fc3-2966-446c-96d1-fa9203b19280-catalog-content\") pod \"community-operators-8755w\" (UID: \"889a8fc3-2966-446c-96d1-fa9203b19280\") " pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.965299 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/889a8fc3-2966-446c-96d1-fa9203b19280-utilities\") pod \"community-operators-8755w\" (UID: \"889a8fc3-2966-446c-96d1-fa9203b19280\") " pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.992788 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6k9f\" (UniqueName: \"kubernetes.io/projected/889a8fc3-2966-446c-96d1-fa9203b19280-kube-api-access-p6k9f\") pod \"community-operators-8755w\" (UID: \"889a8fc3-2966-446c-96d1-fa9203b19280\") " pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:09 crc kubenswrapper[4744]: I0106 14:41:09.996912 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sqlqh"] Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.003943 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.009948 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.011657 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sqlqh"] Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.066057 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df-catalog-content\") pod \"certified-operators-sqlqh\" (UID: \"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df\") " pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.066129 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df-utilities\") pod \"certified-operators-sqlqh\" (UID: \"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df\") " pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.066288 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnbw7\" (UniqueName: \"kubernetes.io/projected/bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df-kube-api-access-tnbw7\") pod \"certified-operators-sqlqh\" (UID: \"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df\") " pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.124367 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.167890 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnbw7\" (UniqueName: \"kubernetes.io/projected/bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df-kube-api-access-tnbw7\") pod \"certified-operators-sqlqh\" (UID: \"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df\") " pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.168235 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df-catalog-content\") pod \"certified-operators-sqlqh\" (UID: \"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df\") " pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.168635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df-utilities\") pod \"certified-operators-sqlqh\" (UID: \"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df\") " pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.168849 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df-catalog-content\") pod \"certified-operators-sqlqh\" (UID: \"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df\") " pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.168879 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df-utilities\") pod \"certified-operators-sqlqh\" (UID: \"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df\") " pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.189729 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnbw7\" (UniqueName: \"kubernetes.io/projected/bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df-kube-api-access-tnbw7\") pod \"certified-operators-sqlqh\" (UID: \"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df\") " pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.415183 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.508727 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pssbh" event={"ID":"d470584a-0f48-436b-9ead-06f8947b7115","Type":"ContainerStarted","Data":"20fdfd1535242841e68a7bb074b0b90c27a54066af05fbc3844881d3065c5c54"} Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.512127 4744 generic.go:334] "Generic (PLEG): container finished" podID="78579813-4567-46c4-8353-4e6418ba7722" containerID="d9839f2c061ad2661413cd17f45ae407deb018f452ed883f8d508e9038720e75" exitCode=0 Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.512187 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrxvs" event={"ID":"78579813-4567-46c4-8353-4e6418ba7722","Type":"ContainerDied","Data":"d9839f2c061ad2661413cd17f45ae407deb018f452ed883f8d508e9038720e75"} Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.531884 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pssbh" podStartSLOduration=2.0907731370000002 podStartE2EDuration="3.531859777s" podCreationTimestamp="2026-01-06 14:41:07 +0000 UTC" firstStartedPulling="2026-01-06 14:41:08.481649181 +0000 UTC m=+265.109115509" lastFinishedPulling="2026-01-06 14:41:09.922735831 +0000 UTC m=+266.550202149" observedRunningTime="2026-01-06 14:41:10.53086012 +0000 UTC m=+267.158326438" watchObservedRunningTime="2026-01-06 14:41:10.531859777 +0000 UTC m=+267.159326105" Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.558008 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8755w"] Jan 06 14:41:10 crc kubenswrapper[4744]: I0106 14:41:10.831588 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sqlqh"] Jan 06 14:41:10 crc kubenswrapper[4744]: W0106 14:41:10.838498 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbbeaa4ea_0334_48cd_9ddd_f7a5ea6ac2df.slice/crio-f4afe0cdf231c32830501692fee5b9c2ff2f2ac3cbbf31e011add58311e0e992 WatchSource:0}: Error finding container f4afe0cdf231c32830501692fee5b9c2ff2f2ac3cbbf31e011add58311e0e992: Status 404 returned error can't find the container with id f4afe0cdf231c32830501692fee5b9c2ff2f2ac3cbbf31e011add58311e0e992 Jan 06 14:41:11 crc kubenswrapper[4744]: I0106 14:41:11.519652 4744 generic.go:334] "Generic (PLEG): container finished" podID="889a8fc3-2966-446c-96d1-fa9203b19280" containerID="de4385b104ba50de4ec1b13abd5cb4db4a0c17cc1aa3de4fdd4c4c3519f42980" exitCode=0 Jan 06 14:41:11 crc kubenswrapper[4744]: I0106 14:41:11.519765 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8755w" event={"ID":"889a8fc3-2966-446c-96d1-fa9203b19280","Type":"ContainerDied","Data":"de4385b104ba50de4ec1b13abd5cb4db4a0c17cc1aa3de4fdd4c4c3519f42980"} Jan 06 14:41:11 crc kubenswrapper[4744]: I0106 14:41:11.519982 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8755w" event={"ID":"889a8fc3-2966-446c-96d1-fa9203b19280","Type":"ContainerStarted","Data":"50d71293a4e5894dde662fd55cbb098442095eaed774a277ba1cd8bf0df76e24"} Jan 06 14:41:11 crc kubenswrapper[4744]: I0106 14:41:11.523329 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vrxvs" event={"ID":"78579813-4567-46c4-8353-4e6418ba7722","Type":"ContainerStarted","Data":"49bba404beebed27f255bf091f250b70e6faa2aaf106bcbae09fa195791d3f77"} Jan 06 14:41:11 crc kubenswrapper[4744]: I0106 14:41:11.525722 4744 generic.go:334] "Generic (PLEG): container finished" podID="bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df" containerID="cfb8a9c8e5ddffc4c99e354d209ee5a570933198c06fa44ca91bc8d5ff8caf7b" exitCode=0 Jan 06 14:41:11 crc kubenswrapper[4744]: I0106 14:41:11.525908 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sqlqh" event={"ID":"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df","Type":"ContainerDied","Data":"cfb8a9c8e5ddffc4c99e354d209ee5a570933198c06fa44ca91bc8d5ff8caf7b"} Jan 06 14:41:11 crc kubenswrapper[4744]: I0106 14:41:11.527489 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sqlqh" event={"ID":"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df","Type":"ContainerStarted","Data":"f4afe0cdf231c32830501692fee5b9c2ff2f2ac3cbbf31e011add58311e0e992"} Jan 06 14:41:11 crc kubenswrapper[4744]: I0106 14:41:11.590235 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vrxvs" podStartSLOduration=2.057041419 podStartE2EDuration="4.590210276s" podCreationTimestamp="2026-01-06 14:41:07 +0000 UTC" firstStartedPulling="2026-01-06 14:41:08.47790572 +0000 UTC m=+265.105372038" lastFinishedPulling="2026-01-06 14:41:11.011074577 +0000 UTC m=+267.638540895" observedRunningTime="2026-01-06 14:41:11.586677271 +0000 UTC m=+268.214143589" watchObservedRunningTime="2026-01-06 14:41:11.590210276 +0000 UTC m=+268.217676594" Jan 06 14:41:12 crc kubenswrapper[4744]: I0106 14:41:12.532645 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sqlqh" event={"ID":"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df","Type":"ContainerStarted","Data":"9fca22fdf8963dce89b7b7e0c4c10605be83677cce7a7b1dacecef11c4ad8209"} Jan 06 14:41:12 crc kubenswrapper[4744]: I0106 14:41:12.599411 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-cvggf" Jan 06 14:41:12 crc kubenswrapper[4744]: I0106 14:41:12.649027 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8ltv4"] Jan 06 14:41:13 crc kubenswrapper[4744]: I0106 14:41:13.539049 4744 generic.go:334] "Generic (PLEG): container finished" podID="889a8fc3-2966-446c-96d1-fa9203b19280" containerID="1b573439c5cd9503c1bb1720cfb873d271b5782eb08b3af03bd232d2781352d9" exitCode=0 Jan 06 14:41:13 crc kubenswrapper[4744]: I0106 14:41:13.539104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8755w" event={"ID":"889a8fc3-2966-446c-96d1-fa9203b19280","Type":"ContainerDied","Data":"1b573439c5cd9503c1bb1720cfb873d271b5782eb08b3af03bd232d2781352d9"} Jan 06 14:41:13 crc kubenswrapper[4744]: I0106 14:41:13.540976 4744 generic.go:334] "Generic (PLEG): container finished" podID="bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df" containerID="9fca22fdf8963dce89b7b7e0c4c10605be83677cce7a7b1dacecef11c4ad8209" exitCode=0 Jan 06 14:41:13 crc kubenswrapper[4744]: I0106 14:41:13.541010 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sqlqh" event={"ID":"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df","Type":"ContainerDied","Data":"9fca22fdf8963dce89b7b7e0c4c10605be83677cce7a7b1dacecef11c4ad8209"} Jan 06 14:41:15 crc kubenswrapper[4744]: I0106 14:41:15.553062 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sqlqh" event={"ID":"bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df","Type":"ContainerStarted","Data":"1604b28cd60b2b41199c7439064ebbe12c57bee0a0defa2ee87c781a0e7d50e8"} Jan 06 14:41:15 crc kubenswrapper[4744]: I0106 14:41:15.557244 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8755w" event={"ID":"889a8fc3-2966-446c-96d1-fa9203b19280","Type":"ContainerStarted","Data":"71a25989760d00e4b0b120fd7acee85af724704609d8bd3abf1861464f173f30"} Jan 06 14:41:15 crc kubenswrapper[4744]: I0106 14:41:15.569938 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sqlqh" podStartSLOduration=3.899678585 podStartE2EDuration="6.569921011s" podCreationTimestamp="2026-01-06 14:41:09 +0000 UTC" firstStartedPulling="2026-01-06 14:41:11.529313257 +0000 UTC m=+268.156779565" lastFinishedPulling="2026-01-06 14:41:14.199555673 +0000 UTC m=+270.827021991" observedRunningTime="2026-01-06 14:41:15.566349385 +0000 UTC m=+272.193815713" watchObservedRunningTime="2026-01-06 14:41:15.569921011 +0000 UTC m=+272.197387329" Jan 06 14:41:15 crc kubenswrapper[4744]: I0106 14:41:15.590197 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8755w" podStartSLOduration=3.830668837 podStartE2EDuration="6.590175896s" podCreationTimestamp="2026-01-06 14:41:09 +0000 UTC" firstStartedPulling="2026-01-06 14:41:11.521518667 +0000 UTC m=+268.148984985" lastFinishedPulling="2026-01-06 14:41:14.281025726 +0000 UTC m=+270.908492044" observedRunningTime="2026-01-06 14:41:15.585941212 +0000 UTC m=+272.213407540" watchObservedRunningTime="2026-01-06 14:41:15.590175896 +0000 UTC m=+272.217642214" Jan 06 14:41:17 crc kubenswrapper[4744]: I0106 14:41:17.726248 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:17 crc kubenswrapper[4744]: I0106 14:41:17.729514 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:17 crc kubenswrapper[4744]: I0106 14:41:17.772204 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:17 crc kubenswrapper[4744]: I0106 14:41:17.921138 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:17 crc kubenswrapper[4744]: I0106 14:41:17.921247 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:17 crc kubenswrapper[4744]: I0106 14:41:17.962298 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:18 crc kubenswrapper[4744]: I0106 14:41:18.613874 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vrxvs" Jan 06 14:41:18 crc kubenswrapper[4744]: I0106 14:41:18.618376 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pssbh" Jan 06 14:41:20 crc kubenswrapper[4744]: I0106 14:41:20.127806 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:20 crc kubenswrapper[4744]: I0106 14:41:20.128066 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:20 crc kubenswrapper[4744]: I0106 14:41:20.192557 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:20 crc kubenswrapper[4744]: I0106 14:41:20.417028 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:20 crc kubenswrapper[4744]: I0106 14:41:20.417079 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:20 crc kubenswrapper[4744]: I0106 14:41:20.466929 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:20 crc kubenswrapper[4744]: I0106 14:41:20.621794 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8755w" Jan 06 14:41:20 crc kubenswrapper[4744]: I0106 14:41:20.622814 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sqlqh" Jan 06 14:41:37 crc kubenswrapper[4744]: I0106 14:41:37.693383 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" podUID="271db951-8bd1-43de-a5bd-6af64a054375" containerName="registry" containerID="cri-o://d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada" gracePeriod=30 Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.140196 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.229803 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"271db951-8bd1-43de-a5bd-6af64a054375\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.229868 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-trusted-ca\") pod \"271db951-8bd1-43de-a5bd-6af64a054375\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.229902 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-registry-certificates\") pod \"271db951-8bd1-43de-a5bd-6af64a054375\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.229953 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-bound-sa-token\") pod \"271db951-8bd1-43de-a5bd-6af64a054375\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.229981 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7jns\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-kube-api-access-d7jns\") pod \"271db951-8bd1-43de-a5bd-6af64a054375\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.230019 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/271db951-8bd1-43de-a5bd-6af64a054375-installation-pull-secrets\") pod \"271db951-8bd1-43de-a5bd-6af64a054375\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.230130 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/271db951-8bd1-43de-a5bd-6af64a054375-ca-trust-extracted\") pod \"271db951-8bd1-43de-a5bd-6af64a054375\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.230219 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-registry-tls\") pod \"271db951-8bd1-43de-a5bd-6af64a054375\" (UID: \"271db951-8bd1-43de-a5bd-6af64a054375\") " Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.231410 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "271db951-8bd1-43de-a5bd-6af64a054375" (UID: "271db951-8bd1-43de-a5bd-6af64a054375"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.231972 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "271db951-8bd1-43de-a5bd-6af64a054375" (UID: "271db951-8bd1-43de-a5bd-6af64a054375"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.238346 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "271db951-8bd1-43de-a5bd-6af64a054375" (UID: "271db951-8bd1-43de-a5bd-6af64a054375"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.239222 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "271db951-8bd1-43de-a5bd-6af64a054375" (UID: "271db951-8bd1-43de-a5bd-6af64a054375"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.241696 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/271db951-8bd1-43de-a5bd-6af64a054375-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "271db951-8bd1-43de-a5bd-6af64a054375" (UID: "271db951-8bd1-43de-a5bd-6af64a054375"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.242470 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-kube-api-access-d7jns" (OuterVolumeSpecName: "kube-api-access-d7jns") pod "271db951-8bd1-43de-a5bd-6af64a054375" (UID: "271db951-8bd1-43de-a5bd-6af64a054375"). InnerVolumeSpecName "kube-api-access-d7jns". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.245096 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "271db951-8bd1-43de-a5bd-6af64a054375" (UID: "271db951-8bd1-43de-a5bd-6af64a054375"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.255275 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/271db951-8bd1-43de-a5bd-6af64a054375-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "271db951-8bd1-43de-a5bd-6af64a054375" (UID: "271db951-8bd1-43de-a5bd-6af64a054375"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.331820 4744 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/271db951-8bd1-43de-a5bd-6af64a054375-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.331869 4744 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/271db951-8bd1-43de-a5bd-6af64a054375-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.331889 4744 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.331907 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.331927 4744 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/271db951-8bd1-43de-a5bd-6af64a054375-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.331945 4744 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.331962 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7jns\" (UniqueName: \"kubernetes.io/projected/271db951-8bd1-43de-a5bd-6af64a054375-kube-api-access-d7jns\") on node \"crc\" DevicePath \"\"" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.684201 4744 generic.go:334] "Generic (PLEG): container finished" podID="271db951-8bd1-43de-a5bd-6af64a054375" containerID="d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada" exitCode=0 Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.684253 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" event={"ID":"271db951-8bd1-43de-a5bd-6af64a054375","Type":"ContainerDied","Data":"d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada"} Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.684281 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.684300 4744 scope.go:117] "RemoveContainer" containerID="d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.684287 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8ltv4" event={"ID":"271db951-8bd1-43de-a5bd-6af64a054375","Type":"ContainerDied","Data":"29b6d1c7ba553bc633153a925b3ced27825bb33e39038f14f38108c0c3046325"} Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.717277 4744 scope.go:117] "RemoveContainer" containerID="d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada" Jan 06 14:41:38 crc kubenswrapper[4744]: E0106 14:41:38.717773 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada\": container with ID starting with d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada not found: ID does not exist" containerID="d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.717806 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada"} err="failed to get container status \"d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada\": rpc error: code = NotFound desc = could not find container \"d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada\": container with ID starting with d76fc0ea5d5a00e80961898c55b4da61f52f2ac32b34c70f991d14ca69d43ada not found: ID does not exist" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.735081 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8ltv4"] Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.742702 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8ltv4"] Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.790070 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2"] Jan 06 14:41:38 crc kubenswrapper[4744]: E0106 14:41:38.790327 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="271db951-8bd1-43de-a5bd-6af64a054375" containerName="registry" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.790340 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="271db951-8bd1-43de-a5bd-6af64a054375" containerName="registry" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.790436 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="271db951-8bd1-43de-a5bd-6af64a054375" containerName="registry" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.790857 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.792882 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.793055 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.793391 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.793514 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.794526 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.802173 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2"] Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.957964 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/13a06df7-3ac1-4072-b223-7cb03e73b79b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-rfgs2\" (UID: \"13a06df7-3ac1-4072-b223-7cb03e73b79b\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.958044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkqxl\" (UniqueName: \"kubernetes.io/projected/13a06df7-3ac1-4072-b223-7cb03e73b79b-kube-api-access-qkqxl\") pod \"cluster-monitoring-operator-6d5b84845-rfgs2\" (UID: \"13a06df7-3ac1-4072-b223-7cb03e73b79b\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:38 crc kubenswrapper[4744]: I0106 14:41:38.958505 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/13a06df7-3ac1-4072-b223-7cb03e73b79b-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-rfgs2\" (UID: \"13a06df7-3ac1-4072-b223-7cb03e73b79b\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.059662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/13a06df7-3ac1-4072-b223-7cb03e73b79b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-rfgs2\" (UID: \"13a06df7-3ac1-4072-b223-7cb03e73b79b\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.059815 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkqxl\" (UniqueName: \"kubernetes.io/projected/13a06df7-3ac1-4072-b223-7cb03e73b79b-kube-api-access-qkqxl\") pod \"cluster-monitoring-operator-6d5b84845-rfgs2\" (UID: \"13a06df7-3ac1-4072-b223-7cb03e73b79b\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.059864 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/13a06df7-3ac1-4072-b223-7cb03e73b79b-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-rfgs2\" (UID: \"13a06df7-3ac1-4072-b223-7cb03e73b79b\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.061903 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/13a06df7-3ac1-4072-b223-7cb03e73b79b-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-rfgs2\" (UID: \"13a06df7-3ac1-4072-b223-7cb03e73b79b\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.079691 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/13a06df7-3ac1-4072-b223-7cb03e73b79b-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-rfgs2\" (UID: \"13a06df7-3ac1-4072-b223-7cb03e73b79b\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.090118 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkqxl\" (UniqueName: \"kubernetes.io/projected/13a06df7-3ac1-4072-b223-7cb03e73b79b-kube-api-access-qkqxl\") pod \"cluster-monitoring-operator-6d5b84845-rfgs2\" (UID: \"13a06df7-3ac1-4072-b223-7cb03e73b79b\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.161535 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.600443 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2"] Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.692587 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" event={"ID":"13a06df7-3ac1-4072-b223-7cb03e73b79b","Type":"ContainerStarted","Data":"aced921e79bad5b301b925b8be6576c4db1894a097e9deb5ec7a12359fd20fa0"} Jan 06 14:41:39 crc kubenswrapper[4744]: I0106 14:41:39.723713 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="271db951-8bd1-43de-a5bd-6af64a054375" path="/var/lib/kubelet/pods/271db951-8bd1-43de-a5bd-6af64a054375/volumes" Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.367684 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5"] Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.368776 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.370503 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.370785 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-6pz8j" Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.388082 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5"] Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.513101 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/c9626172-6c81-40bd-91a7-48e6790e9f7f-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-lmtg5\" (UID: \"c9626172-6c81-40bd-91a7-48e6790e9f7f\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.613981 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/c9626172-6c81-40bd-91a7-48e6790e9f7f-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-lmtg5\" (UID: \"c9626172-6c81-40bd-91a7-48e6790e9f7f\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.621012 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/c9626172-6c81-40bd-91a7-48e6790e9f7f-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-lmtg5\" (UID: \"c9626172-6c81-40bd-91a7-48e6790e9f7f\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.680390 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" Jan 06 14:41:42 crc kubenswrapper[4744]: I0106 14:41:42.715932 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" event={"ID":"13a06df7-3ac1-4072-b223-7cb03e73b79b","Type":"ContainerStarted","Data":"3957ccb3757f34cd19d552279928334d5ec51dd3b9e5f73ce295f14aeaa94d3b"} Jan 06 14:41:43 crc kubenswrapper[4744]: I0106 14:41:43.126485 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-rfgs2" podStartSLOduration=2.9978477740000002 podStartE2EDuration="5.126460381s" podCreationTimestamp="2026-01-06 14:41:38 +0000 UTC" firstStartedPulling="2026-01-06 14:41:39.613734506 +0000 UTC m=+296.241200834" lastFinishedPulling="2026-01-06 14:41:41.742347123 +0000 UTC m=+298.369813441" observedRunningTime="2026-01-06 14:41:42.742772002 +0000 UTC m=+299.370238370" watchObservedRunningTime="2026-01-06 14:41:43.126460381 +0000 UTC m=+299.753926709" Jan 06 14:41:43 crc kubenswrapper[4744]: I0106 14:41:43.126956 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5"] Jan 06 14:41:43 crc kubenswrapper[4744]: W0106 14:41:43.133777 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9626172_6c81_40bd_91a7_48e6790e9f7f.slice/crio-360ef093c5b168ae879505adc81f9a05a701f5fec9f6f1da54d84d05066fd6cf WatchSource:0}: Error finding container 360ef093c5b168ae879505adc81f9a05a701f5fec9f6f1da54d84d05066fd6cf: Status 404 returned error can't find the container with id 360ef093c5b168ae879505adc81f9a05a701f5fec9f6f1da54d84d05066fd6cf Jan 06 14:41:43 crc kubenswrapper[4744]: I0106 14:41:43.543400 4744 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 06 14:41:43 crc kubenswrapper[4744]: I0106 14:41:43.742059 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" event={"ID":"c9626172-6c81-40bd-91a7-48e6790e9f7f","Type":"ContainerStarted","Data":"360ef093c5b168ae879505adc81f9a05a701f5fec9f6f1da54d84d05066fd6cf"} Jan 06 14:41:45 crc kubenswrapper[4744]: I0106 14:41:45.756409 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" event={"ID":"c9626172-6c81-40bd-91a7-48e6790e9f7f","Type":"ContainerStarted","Data":"7c03bba0c8e880d81b199016ff76ace8915f9d16365a4cb19fba823cd670c028"} Jan 06 14:41:45 crc kubenswrapper[4744]: I0106 14:41:45.756791 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" Jan 06 14:41:45 crc kubenswrapper[4744]: I0106 14:41:45.768577 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" Jan 06 14:41:45 crc kubenswrapper[4744]: I0106 14:41:45.779148 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" podStartSLOduration=1.601642161 podStartE2EDuration="3.779126574s" podCreationTimestamp="2026-01-06 14:41:42 +0000 UTC" firstStartedPulling="2026-01-06 14:41:43.137040815 +0000 UTC m=+299.764507133" lastFinishedPulling="2026-01-06 14:41:45.314525238 +0000 UTC m=+301.941991546" observedRunningTime="2026-01-06 14:41:45.778530758 +0000 UTC m=+302.405997076" watchObservedRunningTime="2026-01-06 14:41:45.779126574 +0000 UTC m=+302.406592932" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.442113 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-6cwtj"] Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.444191 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.450019 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-6cwtj"] Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.475462 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.475532 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.475633 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.476203 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-fvmvc" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.476435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2c61e805-8334-4760-83f0-1e4e154d84b8-metrics-client-ca\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.476572 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2c61e805-8334-4760-83f0-1e4e154d84b8-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.476650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6285\" (UniqueName: \"kubernetes.io/projected/2c61e805-8334-4760-83f0-1e4e154d84b8-kube-api-access-r6285\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.476686 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c61e805-8334-4760-83f0-1e4e154d84b8-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.577747 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2c61e805-8334-4760-83f0-1e4e154d84b8-metrics-client-ca\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.577852 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2c61e805-8334-4760-83f0-1e4e154d84b8-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.577900 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6285\" (UniqueName: \"kubernetes.io/projected/2c61e805-8334-4760-83f0-1e4e154d84b8-kube-api-access-r6285\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.577940 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c61e805-8334-4760-83f0-1e4e154d84b8-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: E0106 14:41:46.578113 4744 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-tls: secret "prometheus-operator-tls" not found Jan 06 14:41:46 crc kubenswrapper[4744]: E0106 14:41:46.578213 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c61e805-8334-4760-83f0-1e4e154d84b8-prometheus-operator-tls podName:2c61e805-8334-4760-83f0-1e4e154d84b8 nodeName:}" failed. No retries permitted until 2026-01-06 14:41:47.078188573 +0000 UTC m=+303.705654891 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "prometheus-operator-tls" (UniqueName: "kubernetes.io/secret/2c61e805-8334-4760-83f0-1e4e154d84b8-prometheus-operator-tls") pod "prometheus-operator-db54df47d-6cwtj" (UID: "2c61e805-8334-4760-83f0-1e4e154d84b8") : secret "prometheus-operator-tls" not found Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.579039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2c61e805-8334-4760-83f0-1e4e154d84b8-metrics-client-ca\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.592580 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2c61e805-8334-4760-83f0-1e4e154d84b8-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:46 crc kubenswrapper[4744]: I0106 14:41:46.607015 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6285\" (UniqueName: \"kubernetes.io/projected/2c61e805-8334-4760-83f0-1e4e154d84b8-kube-api-access-r6285\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:47 crc kubenswrapper[4744]: I0106 14:41:47.087096 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c61e805-8334-4760-83f0-1e4e154d84b8-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:47 crc kubenswrapper[4744]: I0106 14:41:47.093352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/2c61e805-8334-4760-83f0-1e4e154d84b8-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-6cwtj\" (UID: \"2c61e805-8334-4760-83f0-1e4e154d84b8\") " pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:47 crc kubenswrapper[4744]: I0106 14:41:47.095017 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" Jan 06 14:41:47 crc kubenswrapper[4744]: I0106 14:41:47.638350 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-6cwtj"] Jan 06 14:41:47 crc kubenswrapper[4744]: W0106 14:41:47.648730 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c61e805_8334_4760_83f0_1e4e154d84b8.slice/crio-4ab17ba4801513e20224c6d217c4b9edbe07a3260b71ec60ae531fd752ee3f3a WatchSource:0}: Error finding container 4ab17ba4801513e20224c6d217c4b9edbe07a3260b71ec60ae531fd752ee3f3a: Status 404 returned error can't find the container with id 4ab17ba4801513e20224c6d217c4b9edbe07a3260b71ec60ae531fd752ee3f3a Jan 06 14:41:47 crc kubenswrapper[4744]: I0106 14:41:47.772672 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" event={"ID":"2c61e805-8334-4760-83f0-1e4e154d84b8","Type":"ContainerStarted","Data":"4ab17ba4801513e20224c6d217c4b9edbe07a3260b71ec60ae531fd752ee3f3a"} Jan 06 14:41:50 crc kubenswrapper[4744]: I0106 14:41:50.801418 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" event={"ID":"2c61e805-8334-4760-83f0-1e4e154d84b8","Type":"ContainerStarted","Data":"cf137e083dcc026726f6a65bd660ba5687538fb53af398091a443b8fcc1e86dc"} Jan 06 14:41:50 crc kubenswrapper[4744]: I0106 14:41:50.802377 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" event={"ID":"2c61e805-8334-4760-83f0-1e4e154d84b8","Type":"ContainerStarted","Data":"6eb0a3a6215e1f31bcd95f4bd1a731703263ad5d423b39986204573d389602ee"} Jan 06 14:41:50 crc kubenswrapper[4744]: I0106 14:41:50.826523 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-6cwtj" podStartSLOduration=2.28324278 podStartE2EDuration="4.826490408s" podCreationTimestamp="2026-01-06 14:41:46 +0000 UTC" firstStartedPulling="2026-01-06 14:41:47.652258005 +0000 UTC m=+304.279724353" lastFinishedPulling="2026-01-06 14:41:50.195505643 +0000 UTC m=+306.822971981" observedRunningTime="2026-01-06 14:41:50.824313419 +0000 UTC m=+307.451779767" watchObservedRunningTime="2026-01-06 14:41:50.826490408 +0000 UTC m=+307.453956816" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.821030 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-grkjv"] Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.822433 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.827897 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.828415 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.829647 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-m2hjj" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.834784 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-grkjv"] Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.845181 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl"] Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.846298 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.851216 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.851288 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-x8gvj" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.854323 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.855639 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.874277 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl"] Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885238 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885304 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885344 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0779f33d-7562-4edc-84b9-8cc87211bf67-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885374 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2z9c\" (UniqueName: \"kubernetes.io/projected/0779f33d-7562-4edc-84b9-8cc87211bf67-kube-api-access-h2z9c\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885425 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp5jl\" (UniqueName: \"kubernetes.io/projected/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-api-access-xp5jl\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885453 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fbbb82b7-a0ef-4e63-85db-16d38444086b-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885480 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/0779f33d-7562-4edc-84b9-8cc87211bf67-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885500 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/fbbb82b7-a0ef-4e63-85db-16d38444086b-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.885522 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/0779f33d-7562-4edc-84b9-8cc87211bf67-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.893703 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-jkwxv"] Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.895059 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.899429 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-wcx67" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.899632 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.899878 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.986915 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fbbb82b7-a0ef-4e63-85db-16d38444086b-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987475 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/c50b38eb-cb07-491b-856b-e0b3edffeadc-root\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987515 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/0779f33d-7562-4edc-84b9-8cc87211bf67-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987543 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/fbbb82b7-a0ef-4e63-85db-16d38444086b-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987570 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-textfile\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987591 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/0779f33d-7562-4edc-84b9-8cc87211bf67-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987613 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987645 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c50b38eb-cb07-491b-856b-e0b3edffeadc-metrics-client-ca\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987673 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987698 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c50b38eb-cb07-491b-856b-e0b3edffeadc-sys\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987718 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-wtmp\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987747 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987772 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-tls\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987803 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0779f33d-7562-4edc-84b9-8cc87211bf67-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987825 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p4wg\" (UniqueName: \"kubernetes.io/projected/c50b38eb-cb07-491b-856b-e0b3edffeadc-kube-api-access-8p4wg\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987855 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2z9c\" (UniqueName: \"kubernetes.io/projected/0779f33d-7562-4edc-84b9-8cc87211bf67-kube-api-access-h2z9c\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987879 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.987900 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp5jl\" (UniqueName: \"kubernetes.io/projected/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-api-access-xp5jl\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.988951 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/fbbb82b7-a0ef-4e63-85db-16d38444086b-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: E0106 14:41:52.989073 4744 secret.go:188] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: secret "openshift-state-metrics-tls" not found Jan 06 14:41:52 crc kubenswrapper[4744]: E0106 14:41:52.989151 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0779f33d-7562-4edc-84b9-8cc87211bf67-openshift-state-metrics-tls podName:0779f33d-7562-4edc-84b9-8cc87211bf67 nodeName:}" failed. No retries permitted until 2026-01-06 14:41:53.489130181 +0000 UTC m=+310.116596499 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/0779f33d-7562-4edc-84b9-8cc87211bf67-openshift-state-metrics-tls") pod "openshift-state-metrics-566fddb674-grkjv" (UID: "0779f33d-7562-4edc-84b9-8cc87211bf67") : secret "openshift-state-metrics-tls" not found Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.989422 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0779f33d-7562-4edc-84b9-8cc87211bf67-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.989507 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.989741 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fbbb82b7-a0ef-4e63-85db-16d38444086b-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.997860 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.998599 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/0779f33d-7562-4edc-84b9-8cc87211bf67-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:52 crc kubenswrapper[4744]: I0106 14:41:52.999643 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.009216 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp5jl\" (UniqueName: \"kubernetes.io/projected/fbbb82b7-a0ef-4e63-85db-16d38444086b-kube-api-access-xp5jl\") pod \"kube-state-metrics-777cb5bd5d-8l7dl\" (UID: \"fbbb82b7-a0ef-4e63-85db-16d38444086b\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.030040 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2z9c\" (UniqueName: \"kubernetes.io/projected/0779f33d-7562-4edc-84b9-8cc87211bf67-kube-api-access-h2z9c\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.088908 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/c50b38eb-cb07-491b-856b-e0b3edffeadc-root\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.088977 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-textfile\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.089003 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.089030 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c50b38eb-cb07-491b-856b-e0b3edffeadc-metrics-client-ca\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.089054 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c50b38eb-cb07-491b-856b-e0b3edffeadc-sys\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.089075 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-wtmp\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.089098 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-tls\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.089126 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p4wg\" (UniqueName: \"kubernetes.io/projected/c50b38eb-cb07-491b-856b-e0b3edffeadc-kube-api-access-8p4wg\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.089460 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/c50b38eb-cb07-491b-856b-e0b3edffeadc-root\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.089794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-textfile\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.090224 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c50b38eb-cb07-491b-856b-e0b3edffeadc-sys\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.090291 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-wtmp\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.091233 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c50b38eb-cb07-491b-856b-e0b3edffeadc-metrics-client-ca\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.094523 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-tls\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.094800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/c50b38eb-cb07-491b-856b-e0b3edffeadc-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.110673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p4wg\" (UniqueName: \"kubernetes.io/projected/c50b38eb-cb07-491b-856b-e0b3edffeadc-kube-api-access-8p4wg\") pod \"node-exporter-jkwxv\" (UID: \"c50b38eb-cb07-491b-856b-e0b3edffeadc\") " pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.158695 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.211349 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-jkwxv" Jan 06 14:41:53 crc kubenswrapper[4744]: W0106 14:41:53.229760 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc50b38eb_cb07_491b_856b_e0b3edffeadc.slice/crio-f23c94040113e3cf62135e7bb13aae68ec6fb91c50d7eefc0f22e85d803f8733 WatchSource:0}: Error finding container f23c94040113e3cf62135e7bb13aae68ec6fb91c50d7eefc0f22e85d803f8733: Status 404 returned error can't find the container with id f23c94040113e3cf62135e7bb13aae68ec6fb91c50d7eefc0f22e85d803f8733 Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.494428 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/0779f33d-7562-4edc-84b9-8cc87211bf67-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.500409 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/0779f33d-7562-4edc-84b9-8cc87211bf67-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-grkjv\" (UID: \"0779f33d-7562-4edc-84b9-8cc87211bf67\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.568704 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl"] Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.737907 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.843254 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-jkwxv" event={"ID":"c50b38eb-cb07-491b-856b-e0b3edffeadc","Type":"ContainerStarted","Data":"f23c94040113e3cf62135e7bb13aae68ec6fb91c50d7eefc0f22e85d803f8733"} Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.848495 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" event={"ID":"fbbb82b7-a0ef-4e63-85db-16d38444086b","Type":"ContainerStarted","Data":"8f05872b141b6d8e83a09be69d7634b3c9461dedd72d0e8ebd76343dab34c29c"} Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.957068 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.981111 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.984319 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.984791 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.985010 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.986173 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.986432 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.986462 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.986659 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.994631 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Jan 06 14:41:53 crc kubenswrapper[4744]: I0106 14:41:53.995864 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-9rzx7" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007153 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/0c248392-1711-4f41-b256-7e5ab7a9baae-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007236 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007464 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007712 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c248392-1711-4f41-b256-7e5ab7a9baae-tls-assets\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007863 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0c248392-1711-4f41-b256-7e5ab7a9baae-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007884 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrj7w\" (UniqueName: \"kubernetes.io/projected/0c248392-1711-4f41-b256-7e5ab7a9baae-kube-api-access-mrj7w\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007914 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-web-config\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007951 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.007997 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c248392-1711-4f41-b256-7e5ab7a9baae-config-out\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.008060 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c248392-1711-4f41-b256-7e5ab7a9baae-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.008135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-config-volume\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.015803 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115251 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c248392-1711-4f41-b256-7e5ab7a9baae-tls-assets\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115390 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0c248392-1711-4f41-b256-7e5ab7a9baae-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115413 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrj7w\" (UniqueName: \"kubernetes.io/projected/0c248392-1711-4f41-b256-7e5ab7a9baae-kube-api-access-mrj7w\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115430 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-web-config\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115449 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115474 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c248392-1711-4f41-b256-7e5ab7a9baae-config-out\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115497 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c248392-1711-4f41-b256-7e5ab7a9baae-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115527 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-config-volume\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115544 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/0c248392-1711-4f41-b256-7e5ab7a9baae-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.115594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.118188 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/0c248392-1711-4f41-b256-7e5ab7a9baae-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.120723 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/0c248392-1711-4f41-b256-7e5ab7a9baae-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.122319 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0c248392-1711-4f41-b256-7e5ab7a9baae-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.135792 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.139072 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.140988 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.145875 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-config-volume\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.145991 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-web-config\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.150984 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0c248392-1711-4f41-b256-7e5ab7a9baae-config-out\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.151542 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0c248392-1711-4f41-b256-7e5ab7a9baae-tls-assets\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.154816 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/0c248392-1711-4f41-b256-7e5ab7a9baae-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.157446 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrj7w\" (UniqueName: \"kubernetes.io/projected/0c248392-1711-4f41-b256-7e5ab7a9baae-kube-api-access-mrj7w\") pod \"alertmanager-main-0\" (UID: \"0c248392-1711-4f41-b256-7e5ab7a9baae\") " pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.211900 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-grkjv"] Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.370424 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.856400 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" event={"ID":"0779f33d-7562-4edc-84b9-8cc87211bf67","Type":"ContainerStarted","Data":"1d81f40d030ad8ae6dcccafe8eddcfcd66a23e32c2b492eb690beaec4b61f65a"} Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.857054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" event={"ID":"0779f33d-7562-4edc-84b9-8cc87211bf67","Type":"ContainerStarted","Data":"f7d4c650c5e988f1b93ce1ef99910555e5921f7d62cc8e5380da1341f9c4cbbd"} Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.857070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" event={"ID":"0779f33d-7562-4edc-84b9-8cc87211bf67","Type":"ContainerStarted","Data":"edccb1b19401a888e0ecee661c09c4558d7da01a840b3d45b5b123c7241f6819"} Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.889073 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Jan 06 14:41:54 crc kubenswrapper[4744]: W0106 14:41:54.895738 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c248392_1711_4f41_b256_7e5ab7a9baae.slice/crio-a95c1dc5d4af3a3f5acfec020b53ba02da1162eb825ea88216d00b9d2624a41e WatchSource:0}: Error finding container a95c1dc5d4af3a3f5acfec020b53ba02da1162eb825ea88216d00b9d2624a41e: Status 404 returned error can't find the container with id a95c1dc5d4af3a3f5acfec020b53ba02da1162eb825ea88216d00b9d2624a41e Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.933178 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-7589f885b-g5hbp"] Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.935082 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.937047 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-47c45fh52m9q8" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.938645 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.938872 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.938954 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.939299 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.939366 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.939547 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-kq7q9" Jan 06 14:41:54 crc kubenswrapper[4744]: I0106 14:41:54.945662 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7589f885b-g5hbp"] Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.032203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.032266 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.032295 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.032445 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjzm6\" (UniqueName: \"kubernetes.io/projected/c69a5e49-5d89-4abc-87fb-ae3749926f17-kube-api-access-sjzm6\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.032662 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-tls\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.032735 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-grpc-tls\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.032828 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c69a5e49-5d89-4abc-87fb-ae3749926f17-metrics-client-ca\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.032865 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.134289 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjzm6\" (UniqueName: \"kubernetes.io/projected/c69a5e49-5d89-4abc-87fb-ae3749926f17-kube-api-access-sjzm6\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.134401 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-tls\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.134435 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-grpc-tls\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.134471 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c69a5e49-5d89-4abc-87fb-ae3749926f17-metrics-client-ca\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.134504 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.134534 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.134584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.134618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.136124 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c69a5e49-5d89-4abc-87fb-ae3749926f17-metrics-client-ca\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.140184 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.141570 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.141895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.141960 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-grpc-tls\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.142412 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.152275 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/c69a5e49-5d89-4abc-87fb-ae3749926f17-secret-thanos-querier-tls\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.152447 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjzm6\" (UniqueName: \"kubernetes.io/projected/c69a5e49-5d89-4abc-87fb-ae3749926f17-kube-api-access-sjzm6\") pod \"thanos-querier-7589f885b-g5hbp\" (UID: \"c69a5e49-5d89-4abc-87fb-ae3749926f17\") " pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.263211 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:41:55 crc kubenswrapper[4744]: I0106 14:41:55.864677 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"0c248392-1711-4f41-b256-7e5ab7a9baae","Type":"ContainerStarted","Data":"a95c1dc5d4af3a3f5acfec020b53ba02da1162eb825ea88216d00b9d2624a41e"} Jan 06 14:41:56 crc kubenswrapper[4744]: I0106 14:41:56.895568 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" event={"ID":"fbbb82b7-a0ef-4e63-85db-16d38444086b","Type":"ContainerStarted","Data":"59c7df9c90a5c1630479b77b467eb2e57d17d9d76b665736e4f46c2d775b99aa"} Jan 06 14:41:56 crc kubenswrapper[4744]: I0106 14:41:56.896097 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" event={"ID":"fbbb82b7-a0ef-4e63-85db-16d38444086b","Type":"ContainerStarted","Data":"7da37d5f1c128633e07de98da63aace5aef71f2d311e68bad74eb00a2e4ba035"} Jan 06 14:41:56 crc kubenswrapper[4744]: I0106 14:41:56.899275 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-jkwxv" event={"ID":"c50b38eb-cb07-491b-856b-e0b3edffeadc","Type":"ContainerStarted","Data":"8111cf62edaf3747f028c8a82d6f7d2664cbdb877fc7d0cb64000e9773e54f1a"} Jan 06 14:41:56 crc kubenswrapper[4744]: I0106 14:41:56.959126 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7589f885b-g5hbp"] Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.689393 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-684db5fdb6-j7lr5"] Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.690110 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.723908 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-684db5fdb6-j7lr5"] Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.773974 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-config\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.774435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-oauth-serving-cert\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.774472 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-serving-cert\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.774498 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-oauth-config\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.774521 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xpnl\" (UniqueName: \"kubernetes.io/projected/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-kube-api-access-2xpnl\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.774585 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-trusted-ca-bundle\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.774641 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-service-ca\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.875903 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-oauth-serving-cert\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.875975 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-serving-cert\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.876019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-oauth-config\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.876051 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xpnl\" (UniqueName: \"kubernetes.io/projected/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-kube-api-access-2xpnl\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.876139 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-trusted-ca-bundle\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.876284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-service-ca\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.876342 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-config\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.877768 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-config\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.877908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-oauth-serving-cert\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.878800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-trusted-ca-bundle\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.879335 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-service-ca\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.884188 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-serving-cert\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.885293 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-oauth-config\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.901832 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xpnl\" (UniqueName: \"kubernetes.io/projected/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-kube-api-access-2xpnl\") pod \"console-684db5fdb6-j7lr5\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.911929 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" event={"ID":"fbbb82b7-a0ef-4e63-85db-16d38444086b","Type":"ContainerStarted","Data":"6b3d8ad8fa809a25208d48ac8345643acde92d569a1c012961c6ad23af7550b5"} Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.915670 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" event={"ID":"c69a5e49-5d89-4abc-87fb-ae3749926f17","Type":"ContainerStarted","Data":"c1ddbf678474f8021994e6972b344eadaeb7575f953e4aef95e8f450b3c85d8b"} Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.917238 4744 generic.go:334] "Generic (PLEG): container finished" podID="c50b38eb-cb07-491b-856b-e0b3edffeadc" containerID="8111cf62edaf3747f028c8a82d6f7d2664cbdb877fc7d0cb64000e9773e54f1a" exitCode=0 Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.917276 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-jkwxv" event={"ID":"c50b38eb-cb07-491b-856b-e0b3edffeadc","Type":"ContainerDied","Data":"8111cf62edaf3747f028c8a82d6f7d2664cbdb877fc7d0cb64000e9773e54f1a"} Jan 06 14:41:57 crc kubenswrapper[4744]: I0106 14:41:57.937360 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-8l7dl" podStartSLOduration=2.975767287 podStartE2EDuration="5.937342816s" podCreationTimestamp="2026-01-06 14:41:52 +0000 UTC" firstStartedPulling="2026-01-06 14:41:53.582478183 +0000 UTC m=+310.209944501" lastFinishedPulling="2026-01-06 14:41:56.544053712 +0000 UTC m=+313.171520030" observedRunningTime="2026-01-06 14:41:57.929974848 +0000 UTC m=+314.557441166" watchObservedRunningTime="2026-01-06 14:41:57.937342816 +0000 UTC m=+314.564809134" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.004288 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.187628 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv"] Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.188822 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.191316 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.191682 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.192276 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.198930 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-de0s84qln3jar" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.200742 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.200976 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-jsqrb" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.209764 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv"] Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.283323 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-audit-log\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.283366 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7r4s\" (UniqueName: \"kubernetes.io/projected/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-kube-api-access-d7r4s\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.283393 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-metrics-server-audit-profiles\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.283484 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-secret-metrics-server-tls\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.283503 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.283543 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-client-ca-bundle\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:58 crc kubenswrapper[4744]: I0106 14:41:58.283568 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-secret-metrics-client-certs\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.385238 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-client-ca-bundle\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.385296 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-secret-metrics-client-certs\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.385357 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-audit-log\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.385374 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7r4s\" (UniqueName: \"kubernetes.io/projected/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-kube-api-access-d7r4s\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.385390 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-metrics-server-audit-profiles\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.385452 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-secret-metrics-server-tls\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.385473 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.385864 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-audit-log\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.386486 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.386869 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-metrics-server-audit-profiles\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.390257 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-secret-metrics-server-tls\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.390372 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-client-ca-bundle\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.390565 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-secret-metrics-client-certs\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.401953 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7r4s\" (UniqueName: \"kubernetes.io/projected/844556c6-9a7d-44c9-aaf0-04f1a4a0497a-kube-api-access-d7r4s\") pod \"metrics-server-7cb5fc74b9-5dnpv\" (UID: \"844556c6-9a7d-44c9-aaf0-04f1a4a0497a\") " pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.511238 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-684db5fdb6-j7lr5"] Jan 06 14:41:59 crc kubenswrapper[4744]: W0106 14:41:58.518541 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9b5b48e_007b_4728_a3ba_bf238a7d36a0.slice/crio-160c47874f1a8b5f14b459e085c30b48153b73f882bc2dcc5d0a8a511e7a92b5 WatchSource:0}: Error finding container 160c47874f1a8b5f14b459e085c30b48153b73f882bc2dcc5d0a8a511e7a92b5: Status 404 returned error can't find the container with id 160c47874f1a8b5f14b459e085c30b48153b73f882bc2dcc5d0a8a511e7a92b5 Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.520357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.627301 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb"] Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.628636 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.632958 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.633036 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.633083 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb"] Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.690542 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/9149e4db-4b7e-4247-ad40-5b2b9738aa38-monitoring-plugin-cert\") pod \"monitoring-plugin-8468c6fcd4-8kpwb\" (UID: \"9149e4db-4b7e-4247-ad40-5b2b9738aa38\") " pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.792534 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/9149e4db-4b7e-4247-ad40-5b2b9738aa38-monitoring-plugin-cert\") pod \"monitoring-plugin-8468c6fcd4-8kpwb\" (UID: \"9149e4db-4b7e-4247-ad40-5b2b9738aa38\") " pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.801263 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/9149e4db-4b7e-4247-ad40-5b2b9738aa38-monitoring-plugin-cert\") pod \"monitoring-plugin-8468c6fcd4-8kpwb\" (UID: \"9149e4db-4b7e-4247-ad40-5b2b9738aa38\") " pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.923853 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-684db5fdb6-j7lr5" event={"ID":"d9b5b48e-007b-4728-a3ba-bf238a7d36a0","Type":"ContainerStarted","Data":"2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a"} Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.923921 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-684db5fdb6-j7lr5" event={"ID":"d9b5b48e-007b-4728-a3ba-bf238a7d36a0","Type":"ContainerStarted","Data":"160c47874f1a8b5f14b459e085c30b48153b73f882bc2dcc5d0a8a511e7a92b5"} Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.927934 4744 generic.go:334] "Generic (PLEG): container finished" podID="0c248392-1711-4f41-b256-7e5ab7a9baae" containerID="3cc73bf00c8772fd36c8331a8c942c9ea62b59ae242c4e1e5a1af4b593bedeab" exitCode=0 Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.927987 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"0c248392-1711-4f41-b256-7e5ab7a9baae","Type":"ContainerDied","Data":"3cc73bf00c8772fd36c8331a8c942c9ea62b59ae242c4e1e5a1af4b593bedeab"} Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.931480 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" event={"ID":"0779f33d-7562-4edc-84b9-8cc87211bf67","Type":"ContainerStarted","Data":"34be007b3345d64078b923bb4b15592c4a72a0be59589839e8e0884d87d27b6c"} Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.934298 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-jkwxv" event={"ID":"c50b38eb-cb07-491b-856b-e0b3edffeadc","Type":"ContainerStarted","Data":"26ac61e424f9d7f7cc508055822d9d1262f226110d6717d2f74183814e388127"} Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.934317 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-jkwxv" event={"ID":"c50b38eb-cb07-491b-856b-e0b3edffeadc","Type":"ContainerStarted","Data":"522487f225677bfe9499c8f4e60e121e7662847989c47ba94f5f38b0c6ac1592"} Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.968542 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-684db5fdb6-j7lr5" podStartSLOduration=1.9685261939999998 podStartE2EDuration="1.968526194s" podCreationTimestamp="2026-01-06 14:41:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:41:58.949918463 +0000 UTC m=+315.577384781" watchObservedRunningTime="2026-01-06 14:41:58.968526194 +0000 UTC m=+315.595992512" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.968734 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-grkjv" podStartSLOduration=3.546396218 podStartE2EDuration="6.968728039s" podCreationTimestamp="2026-01-06 14:41:52 +0000 UTC" firstStartedPulling="2026-01-06 14:41:54.595479581 +0000 UTC m=+311.222945899" lastFinishedPulling="2026-01-06 14:41:58.017811402 +0000 UTC m=+314.645277720" observedRunningTime="2026-01-06 14:41:58.965980415 +0000 UTC m=+315.593446733" watchObservedRunningTime="2026-01-06 14:41:58.968728039 +0000 UTC m=+315.596194357" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.977654 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:58.989538 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-jkwxv" podStartSLOduration=3.6748115649999997 podStartE2EDuration="6.989520209s" podCreationTimestamp="2026-01-06 14:41:52 +0000 UTC" firstStartedPulling="2026-01-06 14:41:53.232367799 +0000 UTC m=+309.859834117" lastFinishedPulling="2026-01-06 14:41:56.547076443 +0000 UTC m=+313.174542761" observedRunningTime="2026-01-06 14:41:58.981153644 +0000 UTC m=+315.608619992" watchObservedRunningTime="2026-01-06 14:41:58.989520209 +0000 UTC m=+315.616986527" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.196326 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.199811 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.202794 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.202842 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-xjs6p" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.203117 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.203185 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.203233 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.203619 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.203949 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-eqaiakt6p6ftc" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.204766 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.205147 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.204773 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.205453 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.206533 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.213954 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.225388 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300481 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-web-config\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300544 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-config\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300563 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300594 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/83a13a72-9958-4f6a-b169-5f6fda738980-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300624 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/83a13a72-9958-4f6a-b169-5f6fda738980-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300648 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300662 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300681 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300706 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb7d8\" (UniqueName: \"kubernetes.io/projected/83a13a72-9958-4f6a-b169-5f6fda738980-kube-api-access-zb7d8\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300724 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300752 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300772 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/83a13a72-9958-4f6a-b169-5f6fda738980-config-out\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300792 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300807 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300821 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300839 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.300856 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.401838 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb7d8\" (UniqueName: \"kubernetes.io/projected/83a13a72-9958-4f6a-b169-5f6fda738980-kube-api-access-zb7d8\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.401891 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.401937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.401963 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/83a13a72-9958-4f6a-b169-5f6fda738980-config-out\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.401988 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402005 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402037 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402089 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-web-config\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402125 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-config\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402142 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/83a13a72-9958-4f6a-b169-5f6fda738980-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402215 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/83a13a72-9958-4f6a-b169-5f6fda738980-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.402241 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.404088 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.404119 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.407136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.407831 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.407885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.408074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.408480 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.409136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.409268 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/83a13a72-9958-4f6a-b169-5f6fda738980-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.409446 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-config\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.409682 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.412531 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-web-config\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.413908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/83a13a72-9958-4f6a-b169-5f6fda738980-config-out\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.414040 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.414631 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/83a13a72-9958-4f6a-b169-5f6fda738980-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.415766 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.415949 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.418324 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/83a13a72-9958-4f6a-b169-5f6fda738980-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.420427 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb7d8\" (UniqueName: \"kubernetes.io/projected/83a13a72-9958-4f6a-b169-5f6fda738980-kube-api-access-zb7d8\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.420806 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/83a13a72-9958-4f6a-b169-5f6fda738980-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"83a13a72-9958-4f6a-b169-5f6fda738980\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.515728 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:41:59 crc kubenswrapper[4744]: W0106 14:41:59.530997 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod844556c6_9a7d_44c9_aaf0_04f1a4a0497a.slice/crio-1bcf9142dd8bd0cd74e5cc9d020a4e73f07bf6c28872b65c368c2158ef76b1ef WatchSource:0}: Error finding container 1bcf9142dd8bd0cd74e5cc9d020a4e73f07bf6c28872b65c368c2158ef76b1ef: Status 404 returned error can't find the container with id 1bcf9142dd8bd0cd74e5cc9d020a4e73f07bf6c28872b65c368c2158ef76b1ef Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.531977 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv"] Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.541553 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb"] Jan 06 14:41:59 crc kubenswrapper[4744]: W0106 14:41:59.542793 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9149e4db_4b7e_4247_ad40_5b2b9738aa38.slice/crio-b10af1fab7eada717eae666fbd681dcad7a6aabffafab9245aabb78787d7e861 WatchSource:0}: Error finding container b10af1fab7eada717eae666fbd681dcad7a6aabffafab9245aabb78787d7e861: Status 404 returned error can't find the container with id b10af1fab7eada717eae666fbd681dcad7a6aabffafab9245aabb78787d7e861 Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.949642 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" event={"ID":"844556c6-9a7d-44c9-aaf0-04f1a4a0497a","Type":"ContainerStarted","Data":"1bcf9142dd8bd0cd74e5cc9d020a4e73f07bf6c28872b65c368c2158ef76b1ef"} Jan 06 14:41:59 crc kubenswrapper[4744]: I0106 14:41:59.957340 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" event={"ID":"9149e4db-4b7e-4247-ad40-5b2b9738aa38","Type":"ContainerStarted","Data":"b10af1fab7eada717eae666fbd681dcad7a6aabffafab9245aabb78787d7e861"} Jan 06 14:42:00 crc kubenswrapper[4744]: I0106 14:42:00.005915 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Jan 06 14:42:00 crc kubenswrapper[4744]: I0106 14:42:00.963616 4744 generic.go:334] "Generic (PLEG): container finished" podID="83a13a72-9958-4f6a-b169-5f6fda738980" containerID="35cc3a4d05782eb20424f529c45b8d9e93be90590538594bcf40af97e927e852" exitCode=0 Jan 06 14:42:00 crc kubenswrapper[4744]: I0106 14:42:00.963658 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"83a13a72-9958-4f6a-b169-5f6fda738980","Type":"ContainerDied","Data":"35cc3a4d05782eb20424f529c45b8d9e93be90590538594bcf40af97e927e852"} Jan 06 14:42:00 crc kubenswrapper[4744]: I0106 14:42:00.963972 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"83a13a72-9958-4f6a-b169-5f6fda738980","Type":"ContainerStarted","Data":"fb70059cc3ea5033566ec133746993f0b35f7fd6de74d1be11ce4baa354fc7f9"} Jan 06 14:42:00 crc kubenswrapper[4744]: I0106 14:42:00.965720 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" event={"ID":"c69a5e49-5d89-4abc-87fb-ae3749926f17","Type":"ContainerStarted","Data":"6081c7c42b20b96f91832958f70e69f2fe19cbe567d2ae4e4351a9732fd78731"} Jan 06 14:42:00 crc kubenswrapper[4744]: I0106 14:42:00.965758 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" event={"ID":"c69a5e49-5d89-4abc-87fb-ae3749926f17","Type":"ContainerStarted","Data":"44f678d9fa8fffe56034773a6b310f8b96fe74b82f29b72c04bd1e3e6c663b08"} Jan 06 14:42:01 crc kubenswrapper[4744]: I0106 14:42:01.985891 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" event={"ID":"c69a5e49-5d89-4abc-87fb-ae3749926f17","Type":"ContainerStarted","Data":"716a381187f79b7286439b24c82209f2eb016f0422c9ed5d8e65bdfe3604fb42"} Jan 06 14:42:02 crc kubenswrapper[4744]: I0106 14:42:02.992295 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"0c248392-1711-4f41-b256-7e5ab7a9baae","Type":"ContainerStarted","Data":"348d067927aa25e88f47a95819adcf89754b59843c439c7d616487279511d9f8"} Jan 06 14:42:03 crc kubenswrapper[4744]: I0106 14:42:02.993874 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" event={"ID":"9149e4db-4b7e-4247-ad40-5b2b9738aa38","Type":"ContainerStarted","Data":"22dac2633c7d433b6f27fe756c910ba032391fa26555d01ab06c7244aae85d79"} Jan 06 14:42:03 crc kubenswrapper[4744]: I0106 14:42:02.994977 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" Jan 06 14:42:03 crc kubenswrapper[4744]: I0106 14:42:02.997258 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" event={"ID":"844556c6-9a7d-44c9-aaf0-04f1a4a0497a","Type":"ContainerStarted","Data":"8fb44164ac10b435d44bd7eba98a6a65bcced9d5777017efea5890c12febc5dd"} Jan 06 14:42:03 crc kubenswrapper[4744]: I0106 14:42:03.001430 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" Jan 06 14:42:03 crc kubenswrapper[4744]: I0106 14:42:03.038539 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" podStartSLOduration=1.835314576 podStartE2EDuration="5.038520439s" podCreationTimestamp="2026-01-06 14:41:58 +0000 UTC" firstStartedPulling="2026-01-06 14:41:59.546731598 +0000 UTC m=+316.174197926" lastFinishedPulling="2026-01-06 14:42:02.749937471 +0000 UTC m=+319.377403789" observedRunningTime="2026-01-06 14:42:03.01293164 +0000 UTC m=+319.640397958" watchObservedRunningTime="2026-01-06 14:42:03.038520439 +0000 UTC m=+319.665986757" Jan 06 14:42:03 crc kubenswrapper[4744]: I0106 14:42:03.042153 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" podStartSLOduration=1.844836063 podStartE2EDuration="5.042139897s" podCreationTimestamp="2026-01-06 14:41:58 +0000 UTC" firstStartedPulling="2026-01-06 14:41:59.533118072 +0000 UTC m=+316.160584410" lastFinishedPulling="2026-01-06 14:42:02.730421926 +0000 UTC m=+319.357888244" observedRunningTime="2026-01-06 14:42:03.03966409 +0000 UTC m=+319.667130408" watchObservedRunningTime="2026-01-06 14:42:03.042139897 +0000 UTC m=+319.669606215" Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.008594 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"0c248392-1711-4f41-b256-7e5ab7a9baae","Type":"ContainerStarted","Data":"3ae54fef1a62579a2e0bd601e87f8adcf1579c4b606bba81f76bd6a8e3cb85c7"} Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.008648 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"0c248392-1711-4f41-b256-7e5ab7a9baae","Type":"ContainerStarted","Data":"c50f7ccf556bd32cf91f417673526c4d597976c598a0a7e7c15f658930437a13"} Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.008662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"0c248392-1711-4f41-b256-7e5ab7a9baae","Type":"ContainerStarted","Data":"ce47ac6cff6da29abf0bef2ac6a21a20723f4f827e3e76ab95fc010488e109b9"} Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.008675 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"0c248392-1711-4f41-b256-7e5ab7a9baae","Type":"ContainerStarted","Data":"42b12f4ed41a3ea049812a9941e64c6f7b258e12a401212b11b356090cd3f3a4"} Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.008686 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"0c248392-1711-4f41-b256-7e5ab7a9baae","Type":"ContainerStarted","Data":"340ff0b9a93cc2d19fe10567f56dda49df360f3302cb46c4013fae9471d6a68b"} Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.013515 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" event={"ID":"c69a5e49-5d89-4abc-87fb-ae3749926f17","Type":"ContainerStarted","Data":"25edd1975d2b79afc338f782efaebdf337eaff4e880c3df6f732c9ea0f4073fb"} Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.013560 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.013575 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" event={"ID":"c69a5e49-5d89-4abc-87fb-ae3749926f17","Type":"ContainerStarted","Data":"c496998054c3ac4ab6e3f04c3290ab20e231d55ebf69503a940ed2af60a8f65f"} Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.013587 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" event={"ID":"c69a5e49-5d89-4abc-87fb-ae3749926f17","Type":"ContainerStarted","Data":"7730e441b2aad365875457f6beff2e7ca5209f061516b1a2c2998bbf248f1eb5"} Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.071344 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=3.242011922 podStartE2EDuration="11.071324129s" podCreationTimestamp="2026-01-06 14:41:53 +0000 UTC" firstStartedPulling="2026-01-06 14:41:54.900686337 +0000 UTC m=+311.528152655" lastFinishedPulling="2026-01-06 14:42:02.729998534 +0000 UTC m=+319.357464862" observedRunningTime="2026-01-06 14:42:04.06354968 +0000 UTC m=+320.691016038" watchObservedRunningTime="2026-01-06 14:42:04.071324129 +0000 UTC m=+320.698790447" Jan 06 14:42:04 crc kubenswrapper[4744]: I0106 14:42:04.107373 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" podStartSLOduration=3.561796808 podStartE2EDuration="10.107335449s" podCreationTimestamp="2026-01-06 14:41:54 +0000 UTC" firstStartedPulling="2026-01-06 14:41:56.966793901 +0000 UTC m=+313.594260219" lastFinishedPulling="2026-01-06 14:42:03.512332542 +0000 UTC m=+320.139798860" observedRunningTime="2026-01-06 14:42:04.102333014 +0000 UTC m=+320.729799342" watchObservedRunningTime="2026-01-06 14:42:04.107335449 +0000 UTC m=+320.734801777" Jan 06 14:42:05 crc kubenswrapper[4744]: I0106 14:42:05.029664 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-7589f885b-g5hbp" Jan 06 14:42:06 crc kubenswrapper[4744]: I0106 14:42:06.030482 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"83a13a72-9958-4f6a-b169-5f6fda738980","Type":"ContainerStarted","Data":"dbe9479b86fae8d7e505bfcb2a1f29a0ba359b6772d5f6d5810eae0b49b80ddd"} Jan 06 14:42:06 crc kubenswrapper[4744]: I0106 14:42:06.030818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"83a13a72-9958-4f6a-b169-5f6fda738980","Type":"ContainerStarted","Data":"86f23ab8d799ea96ba0f9fdb4e873934955e3b4230c8b6d11939ef720131f90b"} Jan 06 14:42:06 crc kubenswrapper[4744]: I0106 14:42:06.030828 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"83a13a72-9958-4f6a-b169-5f6fda738980","Type":"ContainerStarted","Data":"02016b8f4cdfa14d124c39ff6a95c01df815316fd0f834907de1cf2e5a4c2d41"} Jan 06 14:42:06 crc kubenswrapper[4744]: I0106 14:42:06.030837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"83a13a72-9958-4f6a-b169-5f6fda738980","Type":"ContainerStarted","Data":"6d95f31ae635dd6995290d0ffb2f99d3f222da32924689bfb5193de67a401898"} Jan 06 14:42:07 crc kubenswrapper[4744]: I0106 14:42:07.044370 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"83a13a72-9958-4f6a-b169-5f6fda738980","Type":"ContainerStarted","Data":"f43a59b27bd916e1d790dc689c09c149b15090b3d004be95bd54cfb287424755"} Jan 06 14:42:07 crc kubenswrapper[4744]: I0106 14:42:07.044756 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"83a13a72-9958-4f6a-b169-5f6fda738980","Type":"ContainerStarted","Data":"3563c773c74f8077692f8717d2c7fe435ca558b50d8271ca654d0e93ef4fd5df"} Jan 06 14:42:07 crc kubenswrapper[4744]: I0106 14:42:07.098738 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=3.8232825139999997 podStartE2EDuration="8.09872423s" podCreationTimestamp="2026-01-06 14:41:59 +0000 UTC" firstStartedPulling="2026-01-06 14:42:00.967020068 +0000 UTC m=+317.594486436" lastFinishedPulling="2026-01-06 14:42:05.242461794 +0000 UTC m=+321.869928152" observedRunningTime="2026-01-06 14:42:07.097637901 +0000 UTC m=+323.725104219" watchObservedRunningTime="2026-01-06 14:42:07.09872423 +0000 UTC m=+323.726190548" Jan 06 14:42:08 crc kubenswrapper[4744]: I0106 14:42:08.004730 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:42:08 crc kubenswrapper[4744]: I0106 14:42:08.004821 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:42:08 crc kubenswrapper[4744]: I0106 14:42:08.009955 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:42:08 crc kubenswrapper[4744]: I0106 14:42:08.058286 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:42:08 crc kubenswrapper[4744]: I0106 14:42:08.130452 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6pgv4"] Jan 06 14:42:09 crc kubenswrapper[4744]: I0106 14:42:09.517067 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:42:14 crc kubenswrapper[4744]: I0106 14:42:14.424012 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:42:14 crc kubenswrapper[4744]: I0106 14:42:14.424348 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:42:18 crc kubenswrapper[4744]: I0106 14:42:18.523513 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:42:18 crc kubenswrapper[4744]: I0106 14:42:18.523965 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.187133 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-6pgv4" podUID="0640a05a-4a45-4622-8749-88a0621fc348" containerName="console" containerID="cri-o://757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8" gracePeriod=15 Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.590487 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6pgv4_0640a05a-4a45-4622-8749-88a0621fc348/console/0.log" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.590813 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.711810 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-console-config\") pod \"0640a05a-4a45-4622-8749-88a0621fc348\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.711917 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-serving-cert\") pod \"0640a05a-4a45-4622-8749-88a0621fc348\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.712028 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-oauth-config\") pod \"0640a05a-4a45-4622-8749-88a0621fc348\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.712132 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pbv8\" (UniqueName: \"kubernetes.io/projected/0640a05a-4a45-4622-8749-88a0621fc348-kube-api-access-5pbv8\") pod \"0640a05a-4a45-4622-8749-88a0621fc348\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.712257 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-trusted-ca-bundle\") pod \"0640a05a-4a45-4622-8749-88a0621fc348\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.712355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-oauth-serving-cert\") pod \"0640a05a-4a45-4622-8749-88a0621fc348\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.712547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-service-ca\") pod \"0640a05a-4a45-4622-8749-88a0621fc348\" (UID: \"0640a05a-4a45-4622-8749-88a0621fc348\") " Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.712979 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "0640a05a-4a45-4622-8749-88a0621fc348" (UID: "0640a05a-4a45-4622-8749-88a0621fc348"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.712999 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "0640a05a-4a45-4622-8749-88a0621fc348" (UID: "0640a05a-4a45-4622-8749-88a0621fc348"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.713302 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.713355 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.713443 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-service-ca" (OuterVolumeSpecName: "service-ca") pod "0640a05a-4a45-4622-8749-88a0621fc348" (UID: "0640a05a-4a45-4622-8749-88a0621fc348"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.713486 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-console-config" (OuterVolumeSpecName: "console-config") pod "0640a05a-4a45-4622-8749-88a0621fc348" (UID: "0640a05a-4a45-4622-8749-88a0621fc348"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.717688 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0640a05a-4a45-4622-8749-88a0621fc348-kube-api-access-5pbv8" (OuterVolumeSpecName: "kube-api-access-5pbv8") pod "0640a05a-4a45-4622-8749-88a0621fc348" (UID: "0640a05a-4a45-4622-8749-88a0621fc348"). InnerVolumeSpecName "kube-api-access-5pbv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.717723 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "0640a05a-4a45-4622-8749-88a0621fc348" (UID: "0640a05a-4a45-4622-8749-88a0621fc348"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.718085 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "0640a05a-4a45-4622-8749-88a0621fc348" (UID: "0640a05a-4a45-4622-8749-88a0621fc348"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.814202 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.814241 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0640a05a-4a45-4622-8749-88a0621fc348-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.814254 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pbv8\" (UniqueName: \"kubernetes.io/projected/0640a05a-4a45-4622-8749-88a0621fc348-kube-api-access-5pbv8\") on node \"crc\" DevicePath \"\"" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.814269 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-service-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:42:33 crc kubenswrapper[4744]: I0106 14:42:33.814396 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0640a05a-4a45-4622-8749-88a0621fc348-console-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.260474 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-6pgv4_0640a05a-4a45-4622-8749-88a0621fc348/console/0.log" Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.260546 4744 generic.go:334] "Generic (PLEG): container finished" podID="0640a05a-4a45-4622-8749-88a0621fc348" containerID="757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8" exitCode=2 Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.260582 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6pgv4" event={"ID":"0640a05a-4a45-4622-8749-88a0621fc348","Type":"ContainerDied","Data":"757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8"} Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.260633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6pgv4" event={"ID":"0640a05a-4a45-4622-8749-88a0621fc348","Type":"ContainerDied","Data":"e2a447c38ea2d3b299ee47b6b5e6c876a0de965fbd629d388eb0a9a7bdc5f6af"} Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.260646 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6pgv4" Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.260655 4744 scope.go:117] "RemoveContainer" containerID="757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8" Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.283476 4744 scope.go:117] "RemoveContainer" containerID="757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8" Jan 06 14:42:34 crc kubenswrapper[4744]: E0106 14:42:34.283859 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8\": container with ID starting with 757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8 not found: ID does not exist" containerID="757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8" Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.283885 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8"} err="failed to get container status \"757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8\": rpc error: code = NotFound desc = could not find container \"757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8\": container with ID starting with 757ec64d195752b5bbee883e1336f259ca795892fcde918184448c242e5040c8 not found: ID does not exist" Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.292627 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-6pgv4"] Jan 06 14:42:34 crc kubenswrapper[4744]: I0106 14:42:34.298151 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-6pgv4"] Jan 06 14:42:35 crc kubenswrapper[4744]: I0106 14:42:35.720085 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0640a05a-4a45-4622-8749-88a0621fc348" path="/var/lib/kubelet/pods/0640a05a-4a45-4622-8749-88a0621fc348/volumes" Jan 06 14:42:38 crc kubenswrapper[4744]: I0106 14:42:38.527935 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:42:38 crc kubenswrapper[4744]: I0106 14:42:38.532732 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-7cb5fc74b9-5dnpv" Jan 06 14:42:44 crc kubenswrapper[4744]: I0106 14:42:44.424112 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:42:44 crc kubenswrapper[4744]: I0106 14:42:44.424740 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:42:59 crc kubenswrapper[4744]: I0106 14:42:59.516875 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:42:59 crc kubenswrapper[4744]: I0106 14:42:59.553991 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:43:00 crc kubenswrapper[4744]: I0106 14:43:00.503277 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Jan 06 14:43:14 crc kubenswrapper[4744]: I0106 14:43:14.423884 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:43:14 crc kubenswrapper[4744]: I0106 14:43:14.424584 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:43:14 crc kubenswrapper[4744]: I0106 14:43:14.424654 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:43:14 crc kubenswrapper[4744]: I0106 14:43:14.425685 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a62092cf79e8012550863bd594759780ea4401103ffbb749d1491e8c21a56e8a"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 14:43:14 crc kubenswrapper[4744]: I0106 14:43:14.425825 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://a62092cf79e8012550863bd594759780ea4401103ffbb749d1491e8c21a56e8a" gracePeriod=600 Jan 06 14:43:14 crc kubenswrapper[4744]: I0106 14:43:14.572316 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="a62092cf79e8012550863bd594759780ea4401103ffbb749d1491e8c21a56e8a" exitCode=0 Jan 06 14:43:14 crc kubenswrapper[4744]: I0106 14:43:14.572365 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"a62092cf79e8012550863bd594759780ea4401103ffbb749d1491e8c21a56e8a"} Jan 06 14:43:14 crc kubenswrapper[4744]: I0106 14:43:14.572409 4744 scope.go:117] "RemoveContainer" containerID="f0db91145c265a1792dad8ccd75d8c9a1942a698f2bf63e4a4bdd7ae0b74f422" Jan 06 14:43:15 crc kubenswrapper[4744]: I0106 14:43:15.583442 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"c86abf51f54161b2a9bee59a750138794b41f57e70d71c761e86483f896721fe"} Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.636634 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5dddd7d494-hhcc5"] Jan 06 14:43:30 crc kubenswrapper[4744]: E0106 14:43:30.638063 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0640a05a-4a45-4622-8749-88a0621fc348" containerName="console" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.638100 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0640a05a-4a45-4622-8749-88a0621fc348" containerName="console" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.638431 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0640a05a-4a45-4622-8749-88a0621fc348" containerName="console" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.643545 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.665187 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dddd7d494-hhcc5"] Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.697506 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-service-ca\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.697699 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-trusted-ca-bundle\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.697728 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-oauth-serving-cert\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.697773 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-config\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.697818 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-oauth-config\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.697842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-serving-cert\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.697881 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsqln\" (UniqueName: \"kubernetes.io/projected/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-kube-api-access-hsqln\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.799298 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-oauth-config\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.799341 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-serving-cert\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.799383 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsqln\" (UniqueName: \"kubernetes.io/projected/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-kube-api-access-hsqln\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.799420 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-service-ca\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.799470 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-trusted-ca-bundle\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.799486 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-oauth-serving-cert\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.799538 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-config\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.801138 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-config\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.801149 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-service-ca\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.801328 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-oauth-serving-cert\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.801373 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-trusted-ca-bundle\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.808478 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-serving-cert\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.815678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsqln\" (UniqueName: \"kubernetes.io/projected/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-kube-api-access-hsqln\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.818996 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-oauth-config\") pod \"console-5dddd7d494-hhcc5\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:30 crc kubenswrapper[4744]: I0106 14:43:30.975213 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:31 crc kubenswrapper[4744]: I0106 14:43:31.229725 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5dddd7d494-hhcc5"] Jan 06 14:43:31 crc kubenswrapper[4744]: W0106 14:43:31.241192 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf6de0ec_f51e_4ac1_bb55_ea20d2bf8e25.slice/crio-b6a278793ccc099f8e11894e806a9c3d007fceb8db60793e9f0ab1163f8f23ef WatchSource:0}: Error finding container b6a278793ccc099f8e11894e806a9c3d007fceb8db60793e9f0ab1163f8f23ef: Status 404 returned error can't find the container with id b6a278793ccc099f8e11894e806a9c3d007fceb8db60793e9f0ab1163f8f23ef Jan 06 14:43:31 crc kubenswrapper[4744]: I0106 14:43:31.708069 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dddd7d494-hhcc5" event={"ID":"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25","Type":"ContainerStarted","Data":"6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974"} Jan 06 14:43:31 crc kubenswrapper[4744]: I0106 14:43:31.708116 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dddd7d494-hhcc5" event={"ID":"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25","Type":"ContainerStarted","Data":"b6a278793ccc099f8e11894e806a9c3d007fceb8db60793e9f0ab1163f8f23ef"} Jan 06 14:43:31 crc kubenswrapper[4744]: I0106 14:43:31.743260 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5dddd7d494-hhcc5" podStartSLOduration=1.74323717 podStartE2EDuration="1.74323717s" podCreationTimestamp="2026-01-06 14:43:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:43:31.737840686 +0000 UTC m=+408.365307054" watchObservedRunningTime="2026-01-06 14:43:31.74323717 +0000 UTC m=+408.370703528" Jan 06 14:43:40 crc kubenswrapper[4744]: I0106 14:43:40.975821 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:40 crc kubenswrapper[4744]: I0106 14:43:40.976830 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:40 crc kubenswrapper[4744]: I0106 14:43:40.984354 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:41 crc kubenswrapper[4744]: I0106 14:43:41.797532 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:43:41 crc kubenswrapper[4744]: I0106 14:43:41.869507 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-684db5fdb6-j7lr5"] Jan 06 14:44:06 crc kubenswrapper[4744]: I0106 14:44:06.919112 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-684db5fdb6-j7lr5" podUID="d9b5b48e-007b-4728-a3ba-bf238a7d36a0" containerName="console" containerID="cri-o://2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a" gracePeriod=15 Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.347576 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-684db5fdb6-j7lr5_d9b5b48e-007b-4728-a3ba-bf238a7d36a0/console/0.log" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.348129 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.546142 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-trusted-ca-bundle\") pod \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.546290 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-config\") pod \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.546340 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xpnl\" (UniqueName: \"kubernetes.io/projected/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-kube-api-access-2xpnl\") pod \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.546406 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-service-ca\") pod \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.546585 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-oauth-config\") pod \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.546713 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-oauth-serving-cert\") pod \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.546807 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-serving-cert\") pod \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\" (UID: \"d9b5b48e-007b-4728-a3ba-bf238a7d36a0\") " Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.548112 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-service-ca" (OuterVolumeSpecName: "service-ca") pod "d9b5b48e-007b-4728-a3ba-bf238a7d36a0" (UID: "d9b5b48e-007b-4728-a3ba-bf238a7d36a0"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.548321 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d9b5b48e-007b-4728-a3ba-bf238a7d36a0" (UID: "d9b5b48e-007b-4728-a3ba-bf238a7d36a0"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.548358 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-config" (OuterVolumeSpecName: "console-config") pod "d9b5b48e-007b-4728-a3ba-bf238a7d36a0" (UID: "d9b5b48e-007b-4728-a3ba-bf238a7d36a0"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.548365 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d9b5b48e-007b-4728-a3ba-bf238a7d36a0" (UID: "d9b5b48e-007b-4728-a3ba-bf238a7d36a0"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.555990 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-kube-api-access-2xpnl" (OuterVolumeSpecName: "kube-api-access-2xpnl") pod "d9b5b48e-007b-4728-a3ba-bf238a7d36a0" (UID: "d9b5b48e-007b-4728-a3ba-bf238a7d36a0"). InnerVolumeSpecName "kube-api-access-2xpnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.556362 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d9b5b48e-007b-4728-a3ba-bf238a7d36a0" (UID: "d9b5b48e-007b-4728-a3ba-bf238a7d36a0"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.556534 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d9b5b48e-007b-4728-a3ba-bf238a7d36a0" (UID: "d9b5b48e-007b-4728-a3ba-bf238a7d36a0"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.653466 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.653560 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.653588 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.653614 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.653637 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-console-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.653686 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xpnl\" (UniqueName: \"kubernetes.io/projected/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-kube-api-access-2xpnl\") on node \"crc\" DevicePath \"\"" Jan 06 14:44:07 crc kubenswrapper[4744]: I0106 14:44:07.653713 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d9b5b48e-007b-4728-a3ba-bf238a7d36a0-service-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.006589 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-684db5fdb6-j7lr5_d9b5b48e-007b-4728-a3ba-bf238a7d36a0/console/0.log" Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.006691 4744 generic.go:334] "Generic (PLEG): container finished" podID="d9b5b48e-007b-4728-a3ba-bf238a7d36a0" containerID="2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a" exitCode=2 Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.006761 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-684db5fdb6-j7lr5" event={"ID":"d9b5b48e-007b-4728-a3ba-bf238a7d36a0","Type":"ContainerDied","Data":"2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a"} Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.006832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-684db5fdb6-j7lr5" event={"ID":"d9b5b48e-007b-4728-a3ba-bf238a7d36a0","Type":"ContainerDied","Data":"160c47874f1a8b5f14b459e085c30b48153b73f882bc2dcc5d0a8a511e7a92b5"} Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.006842 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-684db5fdb6-j7lr5" Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.006866 4744 scope.go:117] "RemoveContainer" containerID="2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a" Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.035227 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-684db5fdb6-j7lr5"] Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.040606 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-684db5fdb6-j7lr5"] Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.050660 4744 scope.go:117] "RemoveContainer" containerID="2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a" Jan 06 14:44:08 crc kubenswrapper[4744]: E0106 14:44:08.051678 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a\": container with ID starting with 2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a not found: ID does not exist" containerID="2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a" Jan 06 14:44:08 crc kubenswrapper[4744]: I0106 14:44:08.051740 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a"} err="failed to get container status \"2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a\": rpc error: code = NotFound desc = could not find container \"2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a\": container with ID starting with 2df490675bbdbdc29a13174a8d715f33efff0bd93fd459eaee5882c12acd7d9a not found: ID does not exist" Jan 06 14:44:09 crc kubenswrapper[4744]: I0106 14:44:09.722244 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9b5b48e-007b-4728-a3ba-bf238a7d36a0" path="/var/lib/kubelet/pods/d9b5b48e-007b-4728-a3ba-bf238a7d36a0/volumes" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.207593 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z"] Jan 06 14:45:00 crc kubenswrapper[4744]: E0106 14:45:00.208328 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b5b48e-007b-4728-a3ba-bf238a7d36a0" containerName="console" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.208344 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b5b48e-007b-4728-a3ba-bf238a7d36a0" containerName="console" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.208474 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b5b48e-007b-4728-a3ba-bf238a7d36a0" containerName="console" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.208947 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.211746 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.212132 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.237390 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z"] Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.350598 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgd5f\" (UniqueName: \"kubernetes.io/projected/277904fe-5db6-4bfa-82fe-dca1c673e5dc-kube-api-access-qgd5f\") pod \"collect-profiles-29461845-rd62z\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.350788 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/277904fe-5db6-4bfa-82fe-dca1c673e5dc-secret-volume\") pod \"collect-profiles-29461845-rd62z\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.350836 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/277904fe-5db6-4bfa-82fe-dca1c673e5dc-config-volume\") pod \"collect-profiles-29461845-rd62z\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.452266 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/277904fe-5db6-4bfa-82fe-dca1c673e5dc-secret-volume\") pod \"collect-profiles-29461845-rd62z\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.452355 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/277904fe-5db6-4bfa-82fe-dca1c673e5dc-config-volume\") pod \"collect-profiles-29461845-rd62z\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.452467 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgd5f\" (UniqueName: \"kubernetes.io/projected/277904fe-5db6-4bfa-82fe-dca1c673e5dc-kube-api-access-qgd5f\") pod \"collect-profiles-29461845-rd62z\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.453837 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/277904fe-5db6-4bfa-82fe-dca1c673e5dc-config-volume\") pod \"collect-profiles-29461845-rd62z\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.485448 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgd5f\" (UniqueName: \"kubernetes.io/projected/277904fe-5db6-4bfa-82fe-dca1c673e5dc-kube-api-access-qgd5f\") pod \"collect-profiles-29461845-rd62z\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.486089 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/277904fe-5db6-4bfa-82fe-dca1c673e5dc-secret-volume\") pod \"collect-profiles-29461845-rd62z\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.548135 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:00 crc kubenswrapper[4744]: I0106 14:45:00.732956 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z"] Jan 06 14:45:01 crc kubenswrapper[4744]: I0106 14:45:01.453009 4744 generic.go:334] "Generic (PLEG): container finished" podID="277904fe-5db6-4bfa-82fe-dca1c673e5dc" containerID="d50d792cf16292e03b68b9b70c90fb64c21c89dc2b165dfe33694797822614e0" exitCode=0 Jan 06 14:45:01 crc kubenswrapper[4744]: I0106 14:45:01.453150 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" event={"ID":"277904fe-5db6-4bfa-82fe-dca1c673e5dc","Type":"ContainerDied","Data":"d50d792cf16292e03b68b9b70c90fb64c21c89dc2b165dfe33694797822614e0"} Jan 06 14:45:01 crc kubenswrapper[4744]: I0106 14:45:01.453510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" event={"ID":"277904fe-5db6-4bfa-82fe-dca1c673e5dc","Type":"ContainerStarted","Data":"02d5eedac854ab27e2c7872a399cb12812c08850d18053ed79a378bd3c22fb47"} Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.691388 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.788841 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/277904fe-5db6-4bfa-82fe-dca1c673e5dc-secret-volume\") pod \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.788883 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/277904fe-5db6-4bfa-82fe-dca1c673e5dc-config-volume\") pod \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.788931 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgd5f\" (UniqueName: \"kubernetes.io/projected/277904fe-5db6-4bfa-82fe-dca1c673e5dc-kube-api-access-qgd5f\") pod \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\" (UID: \"277904fe-5db6-4bfa-82fe-dca1c673e5dc\") " Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.789761 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/277904fe-5db6-4bfa-82fe-dca1c673e5dc-config-volume" (OuterVolumeSpecName: "config-volume") pod "277904fe-5db6-4bfa-82fe-dca1c673e5dc" (UID: "277904fe-5db6-4bfa-82fe-dca1c673e5dc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.796066 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/277904fe-5db6-4bfa-82fe-dca1c673e5dc-kube-api-access-qgd5f" (OuterVolumeSpecName: "kube-api-access-qgd5f") pod "277904fe-5db6-4bfa-82fe-dca1c673e5dc" (UID: "277904fe-5db6-4bfa-82fe-dca1c673e5dc"). InnerVolumeSpecName "kube-api-access-qgd5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.799998 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277904fe-5db6-4bfa-82fe-dca1c673e5dc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "277904fe-5db6-4bfa-82fe-dca1c673e5dc" (UID: "277904fe-5db6-4bfa-82fe-dca1c673e5dc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.891273 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/277904fe-5db6-4bfa-82fe-dca1c673e5dc-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.891324 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/277904fe-5db6-4bfa-82fe-dca1c673e5dc-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 14:45:02 crc kubenswrapper[4744]: I0106 14:45:02.891345 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgd5f\" (UniqueName: \"kubernetes.io/projected/277904fe-5db6-4bfa-82fe-dca1c673e5dc-kube-api-access-qgd5f\") on node \"crc\" DevicePath \"\"" Jan 06 14:45:03 crc kubenswrapper[4744]: I0106 14:45:03.467374 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" event={"ID":"277904fe-5db6-4bfa-82fe-dca1c673e5dc","Type":"ContainerDied","Data":"02d5eedac854ab27e2c7872a399cb12812c08850d18053ed79a378bd3c22fb47"} Jan 06 14:45:03 crc kubenswrapper[4744]: I0106 14:45:03.467420 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02d5eedac854ab27e2c7872a399cb12812c08850d18053ed79a378bd3c22fb47" Jan 06 14:45:03 crc kubenswrapper[4744]: I0106 14:45:03.467510 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z" Jan 06 14:45:14 crc kubenswrapper[4744]: I0106 14:45:14.424630 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:45:14 crc kubenswrapper[4744]: I0106 14:45:14.426416 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:45:44 crc kubenswrapper[4744]: I0106 14:45:44.424257 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:45:44 crc kubenswrapper[4744]: I0106 14:45:44.425459 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:46:13 crc kubenswrapper[4744]: I0106 14:46:13.804802 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f"] Jan 06 14:46:13 crc kubenswrapper[4744]: E0106 14:46:13.805546 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277904fe-5db6-4bfa-82fe-dca1c673e5dc" containerName="collect-profiles" Jan 06 14:46:13 crc kubenswrapper[4744]: I0106 14:46:13.805560 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="277904fe-5db6-4bfa-82fe-dca1c673e5dc" containerName="collect-profiles" Jan 06 14:46:13 crc kubenswrapper[4744]: I0106 14:46:13.805718 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="277904fe-5db6-4bfa-82fe-dca1c673e5dc" containerName="collect-profiles" Jan 06 14:46:13 crc kubenswrapper[4744]: I0106 14:46:13.806631 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:13 crc kubenswrapper[4744]: I0106 14:46:13.811606 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 06 14:46:13 crc kubenswrapper[4744]: I0106 14:46:13.830011 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f"] Jan 06 14:46:13 crc kubenswrapper[4744]: I0106 14:46:13.917562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:13 crc kubenswrapper[4744]: I0106 14:46:13.917627 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:13 crc kubenswrapper[4744]: I0106 14:46:13.917728 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cfbm\" (UniqueName: \"kubernetes.io/projected/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-kube-api-access-6cfbm\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.019497 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cfbm\" (UniqueName: \"kubernetes.io/projected/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-kube-api-access-6cfbm\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.020276 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.020996 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.023640 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.024644 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.070822 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cfbm\" (UniqueName: \"kubernetes.io/projected/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-kube-api-access-6cfbm\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.140891 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.518802 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.519333 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.519384 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.521413 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c86abf51f54161b2a9bee59a750138794b41f57e70d71c761e86483f896721fe"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.521508 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://c86abf51f54161b2a9bee59a750138794b41f57e70d71c761e86483f896721fe" gracePeriod=600 Jan 06 14:46:14 crc kubenswrapper[4744]: I0106 14:46:14.548397 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f"] Jan 06 14:46:15 crc kubenswrapper[4744]: I0106 14:46:15.006087 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="c86abf51f54161b2a9bee59a750138794b41f57e70d71c761e86483f896721fe" exitCode=0 Jan 06 14:46:15 crc kubenswrapper[4744]: I0106 14:46:15.006123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"c86abf51f54161b2a9bee59a750138794b41f57e70d71c761e86483f896721fe"} Jan 06 14:46:15 crc kubenswrapper[4744]: I0106 14:46:15.006589 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"098e8945831f6c03a5214e0dd27553b61625d63e6339c0ac5bf8762f7f51444d"} Jan 06 14:46:15 crc kubenswrapper[4744]: I0106 14:46:15.006614 4744 scope.go:117] "RemoveContainer" containerID="a62092cf79e8012550863bd594759780ea4401103ffbb749d1491e8c21a56e8a" Jan 06 14:46:15 crc kubenswrapper[4744]: I0106 14:46:15.008893 4744 generic.go:334] "Generic (PLEG): container finished" podID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerID="550f3ccc8da81f659cb1a57e13db768171451b5e5a540871bdebf227fe3a9dfe" exitCode=0 Jan 06 14:46:15 crc kubenswrapper[4744]: I0106 14:46:15.008947 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" event={"ID":"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d","Type":"ContainerDied","Data":"550f3ccc8da81f659cb1a57e13db768171451b5e5a540871bdebf227fe3a9dfe"} Jan 06 14:46:15 crc kubenswrapper[4744]: I0106 14:46:15.008983 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" event={"ID":"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d","Type":"ContainerStarted","Data":"b1140fd088069069f7ca27ba1ba5ee1e5937a4a1678aafdfedd175a693c84632"} Jan 06 14:46:15 crc kubenswrapper[4744]: I0106 14:46:15.010021 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 14:46:17 crc kubenswrapper[4744]: I0106 14:46:17.046959 4744 generic.go:334] "Generic (PLEG): container finished" podID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerID="09c07958a58df5f5218eb65b2a805c83ba79cd07c454a933e82089fe13f408d8" exitCode=0 Jan 06 14:46:17 crc kubenswrapper[4744]: I0106 14:46:17.047086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" event={"ID":"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d","Type":"ContainerDied","Data":"09c07958a58df5f5218eb65b2a805c83ba79cd07c454a933e82089fe13f408d8"} Jan 06 14:46:18 crc kubenswrapper[4744]: I0106 14:46:18.057314 4744 generic.go:334] "Generic (PLEG): container finished" podID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerID="5a21b72c6d70f557f9a1259056a737cf08d6f06fcce157263a6d29ec8365055d" exitCode=0 Jan 06 14:46:18 crc kubenswrapper[4744]: I0106 14:46:18.057436 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" event={"ID":"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d","Type":"ContainerDied","Data":"5a21b72c6d70f557f9a1259056a737cf08d6f06fcce157263a6d29ec8365055d"} Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.387893 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.497507 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-util\") pod \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.497594 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-bundle\") pod \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.497741 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cfbm\" (UniqueName: \"kubernetes.io/projected/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-kube-api-access-6cfbm\") pod \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\" (UID: \"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d\") " Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.501320 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-bundle" (OuterVolumeSpecName: "bundle") pod "e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" (UID: "e99516eb-a6a2-4cf2-b9cd-5f67cf02905d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.506856 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-kube-api-access-6cfbm" (OuterVolumeSpecName: "kube-api-access-6cfbm") pod "e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" (UID: "e99516eb-a6a2-4cf2-b9cd-5f67cf02905d"). InnerVolumeSpecName "kube-api-access-6cfbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.514131 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-util" (OuterVolumeSpecName: "util") pod "e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" (UID: "e99516eb-a6a2-4cf2-b9cd-5f67cf02905d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.599694 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cfbm\" (UniqueName: \"kubernetes.io/projected/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-kube-api-access-6cfbm\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.599766 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-util\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:19 crc kubenswrapper[4744]: I0106 14:46:19.600387 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e99516eb-a6a2-4cf2-b9cd-5f67cf02905d-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:20 crc kubenswrapper[4744]: I0106 14:46:20.077881 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" event={"ID":"e99516eb-a6a2-4cf2-b9cd-5f67cf02905d","Type":"ContainerDied","Data":"b1140fd088069069f7ca27ba1ba5ee1e5937a4a1678aafdfedd175a693c84632"} Jan 06 14:46:20 crc kubenswrapper[4744]: I0106 14:46:20.078322 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1140fd088069069f7ca27ba1ba5ee1e5937a4a1678aafdfedd175a693c84632" Jan 06 14:46:20 crc kubenswrapper[4744]: I0106 14:46:20.077956 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f" Jan 06 14:46:24 crc kubenswrapper[4744]: I0106 14:46:24.746797 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kc5zk"] Jan 06 14:46:24 crc kubenswrapper[4744]: I0106 14:46:24.747534 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovn-controller" containerID="cri-o://feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716" gracePeriod=30 Jan 06 14:46:24 crc kubenswrapper[4744]: I0106 14:46:24.747586 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="sbdb" containerID="cri-o://0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14" gracePeriod=30 Jan 06 14:46:24 crc kubenswrapper[4744]: I0106 14:46:24.747650 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovn-acl-logging" containerID="cri-o://60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603" gracePeriod=30 Jan 06 14:46:24 crc kubenswrapper[4744]: I0106 14:46:24.747648 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="nbdb" containerID="cri-o://dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11" gracePeriod=30 Jan 06 14:46:24 crc kubenswrapper[4744]: I0106 14:46:24.747715 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="northd" containerID="cri-o://30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf" gracePeriod=30 Jan 06 14:46:24 crc kubenswrapper[4744]: I0106 14:46:24.747647 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c" gracePeriod=30 Jan 06 14:46:24 crc kubenswrapper[4744]: I0106 14:46:24.747638 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kube-rbac-proxy-node" containerID="cri-o://e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe" gracePeriod=30 Jan 06 14:46:24 crc kubenswrapper[4744]: I0106 14:46:24.776760 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovnkube-controller" containerID="cri-o://b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39" gracePeriod=30 Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.105993 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jcmms_0cbc5afb-14db-45d6-b564-93a47a9e8123/kube-multus/0.log" Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.106320 4744 generic.go:334] "Generic (PLEG): container finished" podID="0cbc5afb-14db-45d6-b564-93a47a9e8123" containerID="d54efa8efe73152bbe8099225f978ca4b236f6799956aa12b40a9e99d5394847" exitCode=2 Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.106401 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcmms" event={"ID":"0cbc5afb-14db-45d6-b564-93a47a9e8123","Type":"ContainerDied","Data":"d54efa8efe73152bbe8099225f978ca4b236f6799956aa12b40a9e99d5394847"} Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.106896 4744 scope.go:117] "RemoveContainer" containerID="d54efa8efe73152bbe8099225f978ca4b236f6799956aa12b40a9e99d5394847" Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.114629 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kc5zk_9ec2daf2-1809-448d-9891-795e5dff0696/ovn-acl-logging/0.log" Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115114 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kc5zk_9ec2daf2-1809-448d-9891-795e5dff0696/ovn-controller/0.log" Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115449 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ec2daf2-1809-448d-9891-795e5dff0696" containerID="b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39" exitCode=0 Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115470 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ec2daf2-1809-448d-9891-795e5dff0696" containerID="0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14" exitCode=0 Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115478 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ec2daf2-1809-448d-9891-795e5dff0696" containerID="dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11" exitCode=0 Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115485 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ec2daf2-1809-448d-9891-795e5dff0696" containerID="30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf" exitCode=0 Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115491 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ec2daf2-1809-448d-9891-795e5dff0696" containerID="60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603" exitCode=143 Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115499 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ec2daf2-1809-448d-9891-795e5dff0696" containerID="feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716" exitCode=143 Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39"} Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115577 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14"} Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115590 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11"} Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115606 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf"} Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115614 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603"} Jan 06 14:46:25 crc kubenswrapper[4744]: I0106 14:46:25.115623 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716"} Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.045337 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kc5zk_9ec2daf2-1809-448d-9891-795e5dff0696/ovn-acl-logging/0.log" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.045989 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kc5zk_9ec2daf2-1809-448d-9891-795e5dff0696/ovn-controller/0.log" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.046388 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087014 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-etc-openvswitch\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087070 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gp2kl\" (UniqueName: \"kubernetes.io/projected/9ec2daf2-1809-448d-9891-795e5dff0696-kube-api-access-gp2kl\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087090 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-ovn\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087120 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-log-socket\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087135 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-openvswitch\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087148 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-ovn-kubernetes\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087183 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-script-lib\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087198 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087217 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-kubelet\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087237 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-bin\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087257 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-config\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087270 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-env-overrides\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087282 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-node-log\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087301 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9ec2daf2-1809-448d-9891-795e5dff0696-ovn-node-metrics-cert\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087314 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-systemd-units\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087327 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-netns\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087344 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-slash\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087360 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-var-lib-openvswitch\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087375 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-systemd\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087393 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-netd\") pod \"9ec2daf2-1809-448d-9891-795e5dff0696\" (UID: \"9ec2daf2-1809-448d-9891-795e5dff0696\") " Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087607 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.087638 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088139 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088227 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-log-socket" (OuterVolumeSpecName: "log-socket") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088249 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088269 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088368 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-node-log" (OuterVolumeSpecName: "node-log") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088398 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088425 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088446 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088484 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088550 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088570 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-slash" (OuterVolumeSpecName: "host-slash") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088568 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088670 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088760 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.088952 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.106368 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ec2daf2-1809-448d-9891-795e5dff0696-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.112443 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ec2daf2-1809-448d-9891-795e5dff0696-kube-api-access-gp2kl" (OuterVolumeSpecName: "kube-api-access-gp2kl") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "kube-api-access-gp2kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.125669 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9ec2daf2-1809-448d-9891-795e5dff0696" (UID: "9ec2daf2-1809-448d-9891-795e5dff0696"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.131732 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jcmms_0cbc5afb-14db-45d6-b564-93a47a9e8123/kube-multus/0.log" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.131818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jcmms" event={"ID":"0cbc5afb-14db-45d6-b564-93a47a9e8123","Type":"ContainerStarted","Data":"d4828f563defad45aaf5309f4c3add61a4120006736202f29432b91b403ff847"} Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.136447 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kc5zk_9ec2daf2-1809-448d-9891-795e5dff0696/ovn-acl-logging/0.log" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.137774 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-kc5zk_9ec2daf2-1809-448d-9891-795e5dff0696/ovn-controller/0.log" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.141317 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ec2daf2-1809-448d-9891-795e5dff0696" containerID="4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c" exitCode=0 Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.141344 4744 generic.go:334] "Generic (PLEG): container finished" podID="9ec2daf2-1809-448d-9891-795e5dff0696" containerID="e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe" exitCode=0 Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.141363 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c"} Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.141389 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe"} Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.141401 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" event={"ID":"9ec2daf2-1809-448d-9891-795e5dff0696","Type":"ContainerDied","Data":"945ebaac79098f8b6e0344435149a9517e1c4714ebaa58c4f7d970c741f422a8"} Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.141416 4744 scope.go:117] "RemoveContainer" containerID="b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.141548 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-kc5zk" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.172388 4744 scope.go:117] "RemoveContainer" containerID="0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188561 4744 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188590 4744 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-slash\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188599 4744 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188607 4744 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188615 4744 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188623 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gp2kl\" (UniqueName: \"kubernetes.io/projected/9ec2daf2-1809-448d-9891-795e5dff0696-kube-api-access-gp2kl\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188631 4744 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188639 4744 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-log-socket\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188646 4744 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188655 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188664 4744 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188674 4744 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188682 4744 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188690 4744 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188699 4744 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188708 4744 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-node-log\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188715 4744 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9ec2daf2-1809-448d-9891-795e5dff0696-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188723 4744 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188731 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9ec2daf2-1809-448d-9891-795e5dff0696-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.188738 4744 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9ec2daf2-1809-448d-9891-795e5dff0696-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.216438 4744 scope.go:117] "RemoveContainer" containerID="dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.220633 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kc5zk"] Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.240553 4744 scope.go:117] "RemoveContainer" containerID="30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.242688 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-kc5zk"] Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.247869 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bffz5"] Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248080 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovnkube-controller" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248092 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovnkube-controller" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248102 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovn-acl-logging" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248111 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovn-acl-logging" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248123 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerName="pull" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248129 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerName="pull" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248140 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kubecfg-setup" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248146 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kubecfg-setup" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248170 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerName="extract" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248175 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerName="extract" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248185 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="northd" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248198 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="northd" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248205 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="nbdb" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248211 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="nbdb" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248220 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="sbdb" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248228 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="sbdb" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248235 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovn-controller" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248240 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovn-controller" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248249 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerName="util" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248254 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerName="util" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248262 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kube-rbac-proxy-node" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248269 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kube-rbac-proxy-node" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.248275 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kube-rbac-proxy-ovn-metrics" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248281 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kube-rbac-proxy-ovn-metrics" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248375 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kube-rbac-proxy-ovn-metrics" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248386 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovnkube-controller" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248396 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovn-acl-logging" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248402 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="ovn-controller" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248408 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="sbdb" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248414 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="nbdb" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248421 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="northd" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248431 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e99516eb-a6a2-4cf2-b9cd-5f67cf02905d" containerName="extract" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.248438 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" containerName="kube-rbac-proxy-node" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.251041 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.258741 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.258917 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.259072 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.287392 4744 scope.go:117] "RemoveContainer" containerID="4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.289885 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-run-openvswitch\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.289922 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-node-log\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.289945 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.289977 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-etc-openvswitch\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.289999 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-run-netns\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290016 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9691abb2-9584-42d0-88a9-6684011819db-ovn-node-metrics-cert\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290032 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-slash\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290054 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-log-socket\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290071 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-run-ovn-kubernetes\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-cni-bin\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9691abb2-9584-42d0-88a9-6684011819db-ovnkube-script-lib\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290319 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-kubelet\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290336 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-run-ovn\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290354 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9691abb2-9584-42d0-88a9-6684011819db-ovnkube-config\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290393 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-systemd-units\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290411 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-cni-netd\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290437 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnmhd\" (UniqueName: \"kubernetes.io/projected/9691abb2-9584-42d0-88a9-6684011819db-kube-api-access-tnmhd\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290496 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-var-lib-openvswitch\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290525 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9691abb2-9584-42d0-88a9-6684011819db-env-overrides\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.290562 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-run-systemd\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.316972 4744 scope.go:117] "RemoveContainer" containerID="e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.329545 4744 scope.go:117] "RemoveContainer" containerID="60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.344971 4744 scope.go:117] "RemoveContainer" containerID="feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.360872 4744 scope.go:117] "RemoveContainer" containerID="b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.374269 4744 scope.go:117] "RemoveContainer" containerID="b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.374769 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39\": container with ID starting with b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39 not found: ID does not exist" containerID="b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.374813 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39"} err="failed to get container status \"b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39\": rpc error: code = NotFound desc = could not find container \"b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39\": container with ID starting with b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.374841 4744 scope.go:117] "RemoveContainer" containerID="0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.375304 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14\": container with ID starting with 0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14 not found: ID does not exist" containerID="0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.375354 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14"} err="failed to get container status \"0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14\": rpc error: code = NotFound desc = could not find container \"0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14\": container with ID starting with 0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.375382 4744 scope.go:117] "RemoveContainer" containerID="dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.375640 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11\": container with ID starting with dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11 not found: ID does not exist" containerID="dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.375663 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11"} err="failed to get container status \"dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11\": rpc error: code = NotFound desc = could not find container \"dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11\": container with ID starting with dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.375675 4744 scope.go:117] "RemoveContainer" containerID="30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.376009 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf\": container with ID starting with 30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf not found: ID does not exist" containerID="30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.376041 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf"} err="failed to get container status \"30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf\": rpc error: code = NotFound desc = could not find container \"30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf\": container with ID starting with 30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.376054 4744 scope.go:117] "RemoveContainer" containerID="4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.376314 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c\": container with ID starting with 4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c not found: ID does not exist" containerID="4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.376342 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c"} err="failed to get container status \"4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c\": rpc error: code = NotFound desc = could not find container \"4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c\": container with ID starting with 4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.376357 4744 scope.go:117] "RemoveContainer" containerID="e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.376596 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe\": container with ID starting with e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe not found: ID does not exist" containerID="e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.376615 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe"} err="failed to get container status \"e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe\": rpc error: code = NotFound desc = could not find container \"e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe\": container with ID starting with e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.376639 4744 scope.go:117] "RemoveContainer" containerID="60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.376882 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603\": container with ID starting with 60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603 not found: ID does not exist" containerID="60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.376900 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603"} err="failed to get container status \"60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603\": rpc error: code = NotFound desc = could not find container \"60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603\": container with ID starting with 60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.376924 4744 scope.go:117] "RemoveContainer" containerID="feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.377230 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716\": container with ID starting with feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716 not found: ID does not exist" containerID="feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.377250 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716"} err="failed to get container status \"feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716\": rpc error: code = NotFound desc = could not find container \"feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716\": container with ID starting with feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.377262 4744 scope.go:117] "RemoveContainer" containerID="b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789" Jan 06 14:46:26 crc kubenswrapper[4744]: E0106 14:46:26.377500 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789\": container with ID starting with b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789 not found: ID does not exist" containerID="b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.377518 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789"} err="failed to get container status \"b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789\": rpc error: code = NotFound desc = could not find container \"b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789\": container with ID starting with b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.377540 4744 scope.go:117] "RemoveContainer" containerID="b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.377736 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39"} err="failed to get container status \"b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39\": rpc error: code = NotFound desc = could not find container \"b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39\": container with ID starting with b8444be4fe249b1538b039bfb287d39d6c08c6182f114d563c5beb9209d02a39 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.377754 4744 scope.go:117] "RemoveContainer" containerID="0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.377933 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14"} err="failed to get container status \"0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14\": rpc error: code = NotFound desc = could not find container \"0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14\": container with ID starting with 0e63d8f1e3fc7240efea31bfd3ccbb4aa0d38b5803042206a75f77381aabbb14 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.377962 4744 scope.go:117] "RemoveContainer" containerID="dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.378222 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11"} err="failed to get container status \"dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11\": rpc error: code = NotFound desc = could not find container \"dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11\": container with ID starting with dfea648ca304e81f6f2c4bb191eb0b337104986c8ecbc19c27798046a962ca11 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.378254 4744 scope.go:117] "RemoveContainer" containerID="30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.378499 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf"} err="failed to get container status \"30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf\": rpc error: code = NotFound desc = could not find container \"30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf\": container with ID starting with 30a5330ad9613a877e81fcf80c6d5cd50256461ff009adcf8dcd51c46859aecf not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.378516 4744 scope.go:117] "RemoveContainer" containerID="4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.378764 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c"} err="failed to get container status \"4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c\": rpc error: code = NotFound desc = could not find container \"4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c\": container with ID starting with 4d2eb95aea1a665204faccc411e4a9e3af9d72932f3a06f862e3e4662c22f74c not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.378781 4744 scope.go:117] "RemoveContainer" containerID="e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.379035 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe"} err="failed to get container status \"e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe\": rpc error: code = NotFound desc = could not find container \"e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe\": container with ID starting with e61b653ab01dd54591a5d0133a211c4de628db0619565331670d71af4c79febe not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.379062 4744 scope.go:117] "RemoveContainer" containerID="60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.379306 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603"} err="failed to get container status \"60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603\": rpc error: code = NotFound desc = could not find container \"60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603\": container with ID starting with 60fdaeea42677d6dbc0310442519dd4cc2f351315ea3ebf6b894f563ebedb603 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.379334 4744 scope.go:117] "RemoveContainer" containerID="feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.379584 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716"} err="failed to get container status \"feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716\": rpc error: code = NotFound desc = could not find container \"feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716\": container with ID starting with feb145d182aa251f512815d46c1f218cc2cd1c3de07ee2e8d501c1253c717716 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.379612 4744 scope.go:117] "RemoveContainer" containerID="b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.379800 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789"} err="failed to get container status \"b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789\": rpc error: code = NotFound desc = could not find container \"b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789\": container with ID starting with b23da065c4a82a837667fe566531e891e249154c9f61cccee406b3dd29aa6789 not found: ID does not exist" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-etc-openvswitch\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391062 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-run-netns\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391090 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9691abb2-9584-42d0-88a9-6684011819db-ovn-node-metrics-cert\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391116 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-slash\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391145 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-log-socket\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391145 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-etc-openvswitch\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391147 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-run-netns\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391184 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-run-ovn-kubernetes\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391197 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-slash\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-run-ovn-kubernetes\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391387 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-cni-bin\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391406 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-log-socket\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391451 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9691abb2-9584-42d0-88a9-6684011819db-ovnkube-script-lib\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391494 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-kubelet\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391516 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-run-ovn\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391541 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9691abb2-9584-42d0-88a9-6684011819db-ovnkube-config\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391575 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-systemd-units\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391615 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-cni-netd\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnmhd\" (UniqueName: \"kubernetes.io/projected/9691abb2-9584-42d0-88a9-6684011819db-kube-api-access-tnmhd\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391678 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-var-lib-openvswitch\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9691abb2-9584-42d0-88a9-6684011819db-env-overrides\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391726 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-run-systemd\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391825 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-run-openvswitch\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391877 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-systemd-units\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391869 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-node-log\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-cni-netd\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391942 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-var-lib-openvswitch\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.391974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392071 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-node-log\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392098 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392265 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9691abb2-9584-42d0-88a9-6684011819db-ovnkube-config\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392310 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-run-openvswitch\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392319 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-run-systemd\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392331 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-cni-bin\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392341 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-host-kubelet\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392353 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9691abb2-9584-42d0-88a9-6684011819db-run-ovn\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392505 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9691abb2-9584-42d0-88a9-6684011819db-env-overrides\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.392682 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9691abb2-9584-42d0-88a9-6684011819db-ovnkube-script-lib\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.396577 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9691abb2-9584-42d0-88a9-6684011819db-ovn-node-metrics-cert\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.407546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnmhd\" (UniqueName: \"kubernetes.io/projected/9691abb2-9584-42d0-88a9-6684011819db-kube-api-access-tnmhd\") pod \"ovnkube-node-bffz5\" (UID: \"9691abb2-9584-42d0-88a9-6684011819db\") " pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:26 crc kubenswrapper[4744]: I0106 14:46:26.597234 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:27 crc kubenswrapper[4744]: I0106 14:46:27.150050 4744 generic.go:334] "Generic (PLEG): container finished" podID="9691abb2-9584-42d0-88a9-6684011819db" containerID="368b31db5b6d821a55067fcfbac37382546e0c57629c4ad040eb23b65a2066d4" exitCode=0 Jan 06 14:46:27 crc kubenswrapper[4744]: I0106 14:46:27.150217 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerDied","Data":"368b31db5b6d821a55067fcfbac37382546e0c57629c4ad040eb23b65a2066d4"} Jan 06 14:46:27 crc kubenswrapper[4744]: I0106 14:46:27.150399 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerStarted","Data":"d79587263a9cf7cd48c87fe9e88186de96f531060d8e72692d0fec2933db57f7"} Jan 06 14:46:27 crc kubenswrapper[4744]: I0106 14:46:27.717943 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ec2daf2-1809-448d-9891-795e5dff0696" path="/var/lib/kubelet/pods/9ec2daf2-1809-448d-9891-795e5dff0696/volumes" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.160377 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerStarted","Data":"38b3bea95f6f2412304b3858bd4a6110d03deb0b733f86fe7daca6f78b2e5600"} Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.160710 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerStarted","Data":"41e40fd2a616d1175eeaf25eb1527004ecfca7bbfcee9d2143a72e0b04aa3628"} Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.160723 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerStarted","Data":"eb05a25b2a3209fe30eba63cad74c9f379444b81c32af7775153acd708c00623"} Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.160734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerStarted","Data":"da87961051407cbb037a99f0421e261240413be27ff19bc1053334a241b0fce7"} Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.160743 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerStarted","Data":"435392fa9013cb5a19932f5082678ec28c71fd0b188591909fecdf552b113cb7"} Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.160751 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerStarted","Data":"77869394d28db2af73c54266c1be44b5c3b32f1caf17d6e0f0d4ee4a80b1208e"} Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.170201 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-q677r"] Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.171000 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.172413 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.172495 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-sjdws" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.172840 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.216133 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pr9s\" (UniqueName: \"kubernetes.io/projected/7f4652da-367f-462e-92cb-cd4a7acc9487-kube-api-access-9pr9s\") pod \"obo-prometheus-operator-68bc856cb9-q677r\" (UID: \"7f4652da-367f-462e-92cb-cd4a7acc9487\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.295892 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj"] Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.296600 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.300371 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-zrw2f" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.300771 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.321924 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz"] Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.322672 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.324322 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pr9s\" (UniqueName: \"kubernetes.io/projected/7f4652da-367f-462e-92cb-cd4a7acc9487-kube-api-access-9pr9s\") pod \"obo-prometheus-operator-68bc856cb9-q677r\" (UID: \"7f4652da-367f-462e-92cb-cd4a7acc9487\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.324412 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6254c369-a75d-410f-8dc8-1ab47bab3908-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj\" (UID: \"6254c369-a75d-410f-8dc8-1ab47bab3908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.324460 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6254c369-a75d-410f-8dc8-1ab47bab3908-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj\" (UID: \"6254c369-a75d-410f-8dc8-1ab47bab3908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.355607 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pr9s\" (UniqueName: \"kubernetes.io/projected/7f4652da-367f-462e-92cb-cd4a7acc9487-kube-api-access-9pr9s\") pod \"obo-prometheus-operator-68bc856cb9-q677r\" (UID: \"7f4652da-367f-462e-92cb-cd4a7acc9487\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.425415 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e85ae009-0426-4c0e-a1af-de7cd66148cb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz\" (UID: \"e85ae009-0426-4c0e-a1af-de7cd66148cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.425471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e85ae009-0426-4c0e-a1af-de7cd66148cb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz\" (UID: \"e85ae009-0426-4c0e-a1af-de7cd66148cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.425505 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6254c369-a75d-410f-8dc8-1ab47bab3908-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj\" (UID: \"6254c369-a75d-410f-8dc8-1ab47bab3908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.425654 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6254c369-a75d-410f-8dc8-1ab47bab3908-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj\" (UID: \"6254c369-a75d-410f-8dc8-1ab47bab3908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.429283 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6254c369-a75d-410f-8dc8-1ab47bab3908-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj\" (UID: \"6254c369-a75d-410f-8dc8-1ab47bab3908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.429422 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6254c369-a75d-410f-8dc8-1ab47bab3908-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj\" (UID: \"6254c369-a75d-410f-8dc8-1ab47bab3908\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.490126 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.508209 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-cklnq"] Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.509089 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.513942 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.516238 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-q677r_openshift-operators_7f4652da-367f-462e-92cb-cd4a7acc9487_0(03878e2c7143eab0d5d2fdface5290f576b4ddb182bfdc7e5ce4ebbdbe96243b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.516299 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-q677r_openshift-operators_7f4652da-367f-462e-92cb-cd4a7acc9487_0(03878e2c7143eab0d5d2fdface5290f576b4ddb182bfdc7e5ce4ebbdbe96243b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.516323 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-q677r_openshift-operators_7f4652da-367f-462e-92cb-cd4a7acc9487_0(03878e2c7143eab0d5d2fdface5290f576b4ddb182bfdc7e5ce4ebbdbe96243b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.516362 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-68bc856cb9-q677r_openshift-operators(7f4652da-367f-462e-92cb-cd4a7acc9487)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-68bc856cb9-q677r_openshift-operators(7f4652da-367f-462e-92cb-cd4a7acc9487)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-q677r_openshift-operators_7f4652da-367f-462e-92cb-cd4a7acc9487_0(03878e2c7143eab0d5d2fdface5290f576b4ddb182bfdc7e5ce4ebbdbe96243b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" podUID="7f4652da-367f-462e-92cb-cd4a7acc9487" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.516860 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-dx5kb" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.526500 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/313a5ded-3973-4f75-a2ff-d112d168de8d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-cklnq\" (UID: \"313a5ded-3973-4f75-a2ff-d112d168de8d\") " pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.526548 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e85ae009-0426-4c0e-a1af-de7cd66148cb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz\" (UID: \"e85ae009-0426-4c0e-a1af-de7cd66148cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.526611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e85ae009-0426-4c0e-a1af-de7cd66148cb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz\" (UID: \"e85ae009-0426-4c0e-a1af-de7cd66148cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.526650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx7lj\" (UniqueName: \"kubernetes.io/projected/313a5ded-3973-4f75-a2ff-d112d168de8d-kube-api-access-xx7lj\") pod \"observability-operator-59bdc8b94-cklnq\" (UID: \"313a5ded-3973-4f75-a2ff-d112d168de8d\") " pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.529709 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e85ae009-0426-4c0e-a1af-de7cd66148cb-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz\" (UID: \"e85ae009-0426-4c0e-a1af-de7cd66148cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.531609 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e85ae009-0426-4c0e-a1af-de7cd66148cb-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz\" (UID: \"e85ae009-0426-4c0e-a1af-de7cd66148cb\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.599458 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-rx8pm"] Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.600139 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.602906 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-8j6tp" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.609051 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.627832 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx7lj\" (UniqueName: \"kubernetes.io/projected/313a5ded-3973-4f75-a2ff-d112d168de8d-kube-api-access-xx7lj\") pod \"observability-operator-59bdc8b94-cklnq\" (UID: \"313a5ded-3973-4f75-a2ff-d112d168de8d\") " pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.627882 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/81494dce-a93e-4e68-9276-765b36e2cde4-openshift-service-ca\") pod \"perses-operator-5bf474d74f-rx8pm\" (UID: \"81494dce-a93e-4e68-9276-765b36e2cde4\") " pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.627935 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/313a5ded-3973-4f75-a2ff-d112d168de8d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-cklnq\" (UID: \"313a5ded-3973-4f75-a2ff-d112d168de8d\") " pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.627962 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dgps\" (UniqueName: \"kubernetes.io/projected/81494dce-a93e-4e68-9276-765b36e2cde4-kube-api-access-6dgps\") pod \"perses-operator-5bf474d74f-rx8pm\" (UID: \"81494dce-a93e-4e68-9276-765b36e2cde4\") " pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.632730 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/313a5ded-3973-4f75-a2ff-d112d168de8d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-cklnq\" (UID: \"313a5ded-3973-4f75-a2ff-d112d168de8d\") " pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.632837 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators_6254c369-a75d-410f-8dc8-1ab47bab3908_0(989b651c947bc32a18b6a31e41e78ff230513e3e71493ba36d275dcbf42ab8c7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.632898 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators_6254c369-a75d-410f-8dc8-1ab47bab3908_0(989b651c947bc32a18b6a31e41e78ff230513e3e71493ba36d275dcbf42ab8c7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.632924 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators_6254c369-a75d-410f-8dc8-1ab47bab3908_0(989b651c947bc32a18b6a31e41e78ff230513e3e71493ba36d275dcbf42ab8c7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.632973 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators(6254c369-a75d-410f-8dc8-1ab47bab3908)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators(6254c369-a75d-410f-8dc8-1ab47bab3908)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators_6254c369-a75d-410f-8dc8-1ab47bab3908_0(989b651c947bc32a18b6a31e41e78ff230513e3e71493ba36d275dcbf42ab8c7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" podUID="6254c369-a75d-410f-8dc8-1ab47bab3908" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.635123 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.650028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx7lj\" (UniqueName: \"kubernetes.io/projected/313a5ded-3973-4f75-a2ff-d112d168de8d-kube-api-access-xx7lj\") pod \"observability-operator-59bdc8b94-cklnq\" (UID: \"313a5ded-3973-4f75-a2ff-d112d168de8d\") " pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.656775 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators_e85ae009-0426-4c0e-a1af-de7cd66148cb_0(86dbac8e0ce3a4906fbdd2574967c3211e66af72d61e976348ed920293b5f228): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.656841 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators_e85ae009-0426-4c0e-a1af-de7cd66148cb_0(86dbac8e0ce3a4906fbdd2574967c3211e66af72d61e976348ed920293b5f228): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.656881 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators_e85ae009-0426-4c0e-a1af-de7cd66148cb_0(86dbac8e0ce3a4906fbdd2574967c3211e66af72d61e976348ed920293b5f228): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.656934 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators(e85ae009-0426-4c0e-a1af-de7cd66148cb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators(e85ae009-0426-4c0e-a1af-de7cd66148cb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators_e85ae009-0426-4c0e-a1af-de7cd66148cb_0(86dbac8e0ce3a4906fbdd2574967c3211e66af72d61e976348ed920293b5f228): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" podUID="e85ae009-0426-4c0e-a1af-de7cd66148cb" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.729413 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/81494dce-a93e-4e68-9276-765b36e2cde4-openshift-service-ca\") pod \"perses-operator-5bf474d74f-rx8pm\" (UID: \"81494dce-a93e-4e68-9276-765b36e2cde4\") " pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.729758 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dgps\" (UniqueName: \"kubernetes.io/projected/81494dce-a93e-4e68-9276-765b36e2cde4-kube-api-access-6dgps\") pod \"perses-operator-5bf474d74f-rx8pm\" (UID: \"81494dce-a93e-4e68-9276-765b36e2cde4\") " pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.730580 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/81494dce-a93e-4e68-9276-765b36e2cde4-openshift-service-ca\") pod \"perses-operator-5bf474d74f-rx8pm\" (UID: \"81494dce-a93e-4e68-9276-765b36e2cde4\") " pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.751869 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dgps\" (UniqueName: \"kubernetes.io/projected/81494dce-a93e-4e68-9276-765b36e2cde4-kube-api-access-6dgps\") pod \"perses-operator-5bf474d74f-rx8pm\" (UID: \"81494dce-a93e-4e68-9276-765b36e2cde4\") " pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.834320 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.852910 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-cklnq_openshift-operators_313a5ded-3973-4f75-a2ff-d112d168de8d_0(cdf052966860d52a5004aa8d18ada913f5c0106556b4cdb7bd7e870bb95bde5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.853077 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-cklnq_openshift-operators_313a5ded-3973-4f75-a2ff-d112d168de8d_0(cdf052966860d52a5004aa8d18ada913f5c0106556b4cdb7bd7e870bb95bde5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.853195 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-cklnq_openshift-operators_313a5ded-3973-4f75-a2ff-d112d168de8d_0(cdf052966860d52a5004aa8d18ada913f5c0106556b4cdb7bd7e870bb95bde5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.853332 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-59bdc8b94-cklnq_openshift-operators(313a5ded-3973-4f75-a2ff-d112d168de8d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-59bdc8b94-cklnq_openshift-operators(313a5ded-3973-4f75-a2ff-d112d168de8d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-cklnq_openshift-operators_313a5ded-3973-4f75-a2ff-d112d168de8d_0(cdf052966860d52a5004aa8d18ada913f5c0106556b4cdb7bd7e870bb95bde5b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" podUID="313a5ded-3973-4f75-a2ff-d112d168de8d" Jan 06 14:46:28 crc kubenswrapper[4744]: I0106 14:46:28.916319 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.943619 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-rx8pm_openshift-operators_81494dce-a93e-4e68-9276-765b36e2cde4_0(71c8650ef0790a29044c924808004e9bb0ab2fa8fe68f3050cf791efafdddc37): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.943820 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-rx8pm_openshift-operators_81494dce-a93e-4e68-9276-765b36e2cde4_0(71c8650ef0790a29044c924808004e9bb0ab2fa8fe68f3050cf791efafdddc37): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.943913 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-rx8pm_openshift-operators_81494dce-a93e-4e68-9276-765b36e2cde4_0(71c8650ef0790a29044c924808004e9bb0ab2fa8fe68f3050cf791efafdddc37): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:28 crc kubenswrapper[4744]: E0106 14:46:28.944040 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5bf474d74f-rx8pm_openshift-operators(81494dce-a93e-4e68-9276-765b36e2cde4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5bf474d74f-rx8pm_openshift-operators(81494dce-a93e-4e68-9276-765b36e2cde4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-rx8pm_openshift-operators_81494dce-a93e-4e68-9276-765b36e2cde4_0(71c8650ef0790a29044c924808004e9bb0ab2fa8fe68f3050cf791efafdddc37): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" podUID="81494dce-a93e-4e68-9276-765b36e2cde4" Jan 06 14:46:30 crc kubenswrapper[4744]: I0106 14:46:30.179647 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerStarted","Data":"301d90f9245d7499f3653e32437f2668f130c8166da3cd55213788ed8c37ab9c"} Jan 06 14:46:33 crc kubenswrapper[4744]: I0106 14:46:33.208374 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" event={"ID":"9691abb2-9584-42d0-88a9-6684011819db","Type":"ContainerStarted","Data":"3867d0c846f4a88680f51d41e498e5a6a1530b49599f3f1f959f416ece463d06"} Jan 06 14:46:33 crc kubenswrapper[4744]: I0106 14:46:33.209326 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:33 crc kubenswrapper[4744]: I0106 14:46:33.209352 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:33 crc kubenswrapper[4744]: I0106 14:46:33.255686 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" podStartSLOduration=7.255669829 podStartE2EDuration="7.255669829s" podCreationTimestamp="2026-01-06 14:46:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:46:33.249680789 +0000 UTC m=+589.877147117" watchObservedRunningTime="2026-01-06 14:46:33.255669829 +0000 UTC m=+589.883136157" Jan 06 14:46:33 crc kubenswrapper[4744]: I0106 14:46:33.279402 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.054850 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj"] Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.055551 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.056122 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.076376 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz"] Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.076517 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.077030 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.083748 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-cklnq"] Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.083984 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.084679 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.088704 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-q677r"] Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.088836 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.089245 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.101399 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators_6254c369-a75d-410f-8dc8-1ab47bab3908_0(297e164dfff65d271d3c952265343d1ecc20825ae3ded49581d25cd7c01ab607): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.101449 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators_6254c369-a75d-410f-8dc8-1ab47bab3908_0(297e164dfff65d271d3c952265343d1ecc20825ae3ded49581d25cd7c01ab607): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.101468 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators_6254c369-a75d-410f-8dc8-1ab47bab3908_0(297e164dfff65d271d3c952265343d1ecc20825ae3ded49581d25cd7c01ab607): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.101505 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators(6254c369-a75d-410f-8dc8-1ab47bab3908)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators(6254c369-a75d-410f-8dc8-1ab47bab3908)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_openshift-operators_6254c369-a75d-410f-8dc8-1ab47bab3908_0(297e164dfff65d271d3c952265343d1ecc20825ae3ded49581d25cd7c01ab607): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" podUID="6254c369-a75d-410f-8dc8-1ab47bab3908" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.131243 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-rx8pm"] Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.131359 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.131377 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators_e85ae009-0426-4c0e-a1af-de7cd66148cb_0(08a8a9254709c62a13c20f66d2a460b8e1d9c76901fff6ef0ec2313831d11597): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.131435 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators_e85ae009-0426-4c0e-a1af-de7cd66148cb_0(08a8a9254709c62a13c20f66d2a460b8e1d9c76901fff6ef0ec2313831d11597): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.131455 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators_e85ae009-0426-4c0e-a1af-de7cd66148cb_0(08a8a9254709c62a13c20f66d2a460b8e1d9c76901fff6ef0ec2313831d11597): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.131496 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators(e85ae009-0426-4c0e-a1af-de7cd66148cb)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators(e85ae009-0426-4c0e-a1af-de7cd66148cb)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_openshift-operators_e85ae009-0426-4c0e-a1af-de7cd66148cb_0(08a8a9254709c62a13c20f66d2a460b8e1d9c76901fff6ef0ec2313831d11597): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" podUID="e85ae009-0426-4c0e-a1af-de7cd66148cb" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.131739 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.165528 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-cklnq_openshift-operators_313a5ded-3973-4f75-a2ff-d112d168de8d_0(07776ce79d332878aedbe1c617c6f2262e57782dbadf89d5bfd89655760fa830): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.165646 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-cklnq_openshift-operators_313a5ded-3973-4f75-a2ff-d112d168de8d_0(07776ce79d332878aedbe1c617c6f2262e57782dbadf89d5bfd89655760fa830): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.165713 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-cklnq_openshift-operators_313a5ded-3973-4f75-a2ff-d112d168de8d_0(07776ce79d332878aedbe1c617c6f2262e57782dbadf89d5bfd89655760fa830): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.165794 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-59bdc8b94-cklnq_openshift-operators(313a5ded-3973-4f75-a2ff-d112d168de8d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-59bdc8b94-cklnq_openshift-operators(313a5ded-3973-4f75-a2ff-d112d168de8d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-cklnq_openshift-operators_313a5ded-3973-4f75-a2ff-d112d168de8d_0(07776ce79d332878aedbe1c617c6f2262e57782dbadf89d5bfd89655760fa830): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" podUID="313a5ded-3973-4f75-a2ff-d112d168de8d" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.202894 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-q677r_openshift-operators_7f4652da-367f-462e-92cb-cd4a7acc9487_0(e26ce4e78021860a59dc527f3ff0f6186038d59ed9f8fa72e17e3b0895838513): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.202957 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-q677r_openshift-operators_7f4652da-367f-462e-92cb-cd4a7acc9487_0(e26ce4e78021860a59dc527f3ff0f6186038d59ed9f8fa72e17e3b0895838513): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.202976 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-q677r_openshift-operators_7f4652da-367f-462e-92cb-cd4a7acc9487_0(e26ce4e78021860a59dc527f3ff0f6186038d59ed9f8fa72e17e3b0895838513): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.203022 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-68bc856cb9-q677r_openshift-operators(7f4652da-367f-462e-92cb-cd4a7acc9487)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-68bc856cb9-q677r_openshift-operators(7f4652da-367f-462e-92cb-cd4a7acc9487)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-q677r_openshift-operators_7f4652da-367f-462e-92cb-cd4a7acc9487_0(e26ce4e78021860a59dc527f3ff0f6186038d59ed9f8fa72e17e3b0895838513): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" podUID="7f4652da-367f-462e-92cb-cd4a7acc9487" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.210093 4744 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-rx8pm_openshift-operators_81494dce-a93e-4e68-9276-765b36e2cde4_0(dc673c0899efd7d4535b6530e613d5e76d284784f757671675688e6a2e1a91f0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.210139 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-rx8pm_openshift-operators_81494dce-a93e-4e68-9276-765b36e2cde4_0(dc673c0899efd7d4535b6530e613d5e76d284784f757671675688e6a2e1a91f0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.210174 4744 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-rx8pm_openshift-operators_81494dce-a93e-4e68-9276-765b36e2cde4_0(dc673c0899efd7d4535b6530e613d5e76d284784f757671675688e6a2e1a91f0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:34 crc kubenswrapper[4744]: E0106 14:46:34.210209 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5bf474d74f-rx8pm_openshift-operators(81494dce-a93e-4e68-9276-765b36e2cde4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5bf474d74f-rx8pm_openshift-operators(81494dce-a93e-4e68-9276-765b36e2cde4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-rx8pm_openshift-operators_81494dce-a93e-4e68-9276-765b36e2cde4_0(dc673c0899efd7d4535b6530e613d5e76d284784f757671675688e6a2e1a91f0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" podUID="81494dce-a93e-4e68-9276-765b36e2cde4" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.218621 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:34 crc kubenswrapper[4744]: I0106 14:46:34.247762 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:44 crc kubenswrapper[4744]: I0106 14:46:44.710064 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:44 crc kubenswrapper[4744]: I0106 14:46:44.710902 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" Jan 06 14:46:44 crc kubenswrapper[4744]: I0106 14:46:44.971409 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz"] Jan 06 14:46:44 crc kubenswrapper[4744]: W0106 14:46:44.977368 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode85ae009_0426_4c0e_a1af_de7cd66148cb.slice/crio-70ce15542669fd27b1edf5f9c25df8c5600ad2aef54f4044d5874e853aa2bd92 WatchSource:0}: Error finding container 70ce15542669fd27b1edf5f9c25df8c5600ad2aef54f4044d5874e853aa2bd92: Status 404 returned error can't find the container with id 70ce15542669fd27b1edf5f9c25df8c5600ad2aef54f4044d5874e853aa2bd92 Jan 06 14:46:45 crc kubenswrapper[4744]: I0106 14:46:45.284255 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" event={"ID":"e85ae009-0426-4c0e-a1af-de7cd66148cb","Type":"ContainerStarted","Data":"70ce15542669fd27b1edf5f9c25df8c5600ad2aef54f4044d5874e853aa2bd92"} Jan 06 14:46:45 crc kubenswrapper[4744]: I0106 14:46:45.710920 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:45 crc kubenswrapper[4744]: I0106 14:46:45.711023 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:45 crc kubenswrapper[4744]: I0106 14:46:45.711562 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:45 crc kubenswrapper[4744]: I0106 14:46:45.711622 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" Jan 06 14:46:45 crc kubenswrapper[4744]: I0106 14:46:45.995555 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj"] Jan 06 14:46:46 crc kubenswrapper[4744]: W0106 14:46:46.003122 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6254c369_a75d_410f_8dc8_1ab47bab3908.slice/crio-320970c0a2d2c734c576bcb82ba4fd60c3a88aa1aa435966d4cc573db39cf8db WatchSource:0}: Error finding container 320970c0a2d2c734c576bcb82ba4fd60c3a88aa1aa435966d4cc573db39cf8db: Status 404 returned error can't find the container with id 320970c0a2d2c734c576bcb82ba4fd60c3a88aa1aa435966d4cc573db39cf8db Jan 06 14:46:46 crc kubenswrapper[4744]: I0106 14:46:46.143504 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-rx8pm"] Jan 06 14:46:46 crc kubenswrapper[4744]: W0106 14:46:46.147256 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81494dce_a93e_4e68_9276_765b36e2cde4.slice/crio-8b3769ad176d2dcbe6611a99346f6a255f43fae70aea01d149d6a38e6af32f86 WatchSource:0}: Error finding container 8b3769ad176d2dcbe6611a99346f6a255f43fae70aea01d149d6a38e6af32f86: Status 404 returned error can't find the container with id 8b3769ad176d2dcbe6611a99346f6a255f43fae70aea01d149d6a38e6af32f86 Jan 06 14:46:46 crc kubenswrapper[4744]: I0106 14:46:46.291355 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" event={"ID":"6254c369-a75d-410f-8dc8-1ab47bab3908","Type":"ContainerStarted","Data":"320970c0a2d2c734c576bcb82ba4fd60c3a88aa1aa435966d4cc573db39cf8db"} Jan 06 14:46:46 crc kubenswrapper[4744]: I0106 14:46:46.292924 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" event={"ID":"81494dce-a93e-4e68-9276-765b36e2cde4","Type":"ContainerStarted","Data":"8b3769ad176d2dcbe6611a99346f6a255f43fae70aea01d149d6a38e6af32f86"} Jan 06 14:46:47 crc kubenswrapper[4744]: I0106 14:46:47.711476 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:47 crc kubenswrapper[4744]: I0106 14:46:47.712025 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:48 crc kubenswrapper[4744]: I0106 14:46:48.710747 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:48 crc kubenswrapper[4744]: I0106 14:46:48.711413 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.317383 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-cklnq"] Jan 06 14:46:51 crc kubenswrapper[4744]: W0106 14:46:51.319866 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod313a5ded_3973_4f75_a2ff_d112d168de8d.slice/crio-cbcd099917370d6617ff33d5cf8d85b2781ada7da47db20979f212d5ac6b4b92 WatchSource:0}: Error finding container cbcd099917370d6617ff33d5cf8d85b2781ada7da47db20979f212d5ac6b4b92: Status 404 returned error can't find the container with id cbcd099917370d6617ff33d5cf8d85b2781ada7da47db20979f212d5ac6b4b92 Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.333305 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" event={"ID":"81494dce-a93e-4e68-9276-765b36e2cde4","Type":"ContainerStarted","Data":"65163bcf0e99afda387799b6bee12cd5e75199ec506d82f1a9b02464931e7bc6"} Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.333394 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.335337 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" event={"ID":"6254c369-a75d-410f-8dc8-1ab47bab3908","Type":"ContainerStarted","Data":"83e24a12f2822e675ede11b29a7a275ffaad2a079ea4b4bab2d0204f9dd8a9a3"} Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.336331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" event={"ID":"313a5ded-3973-4f75-a2ff-d112d168de8d","Type":"ContainerStarted","Data":"cbcd099917370d6617ff33d5cf8d85b2781ada7da47db20979f212d5ac6b4b92"} Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.337504 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" event={"ID":"e85ae009-0426-4c0e-a1af-de7cd66148cb","Type":"ContainerStarted","Data":"53ead5818fb170d74cc71e6b4bf88dc1c34f984c53af9dac7943f64f1e5faa6c"} Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.381326 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" podStartSLOduration=18.712071712 podStartE2EDuration="23.381310475s" podCreationTimestamp="2026-01-06 14:46:28 +0000 UTC" firstStartedPulling="2026-01-06 14:46:46.150078731 +0000 UTC m=+602.777545049" lastFinishedPulling="2026-01-06 14:46:50.819317494 +0000 UTC m=+607.446783812" observedRunningTime="2026-01-06 14:46:51.379059185 +0000 UTC m=+608.006525503" watchObservedRunningTime="2026-01-06 14:46:51.381310475 +0000 UTC m=+608.008776793" Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.400133 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-q677r"] Jan 06 14:46:51 crc kubenswrapper[4744]: W0106 14:46:51.403639 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f4652da_367f_462e_92cb_cd4a7acc9487.slice/crio-ec56f0c627fd80df7277c1f580eb5762233c823011d9c22db8a225da4e7097c0 WatchSource:0}: Error finding container ec56f0c627fd80df7277c1f580eb5762233c823011d9c22db8a225da4e7097c0: Status 404 returned error can't find the container with id ec56f0c627fd80df7277c1f580eb5762233c823011d9c22db8a225da4e7097c0 Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.421079 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj" podStartSLOduration=18.608319432000002 podStartE2EDuration="23.421065996s" podCreationTimestamp="2026-01-06 14:46:28 +0000 UTC" firstStartedPulling="2026-01-06 14:46:46.005517442 +0000 UTC m=+602.632983760" lastFinishedPulling="2026-01-06 14:46:50.818264006 +0000 UTC m=+607.445730324" observedRunningTime="2026-01-06 14:46:51.419976147 +0000 UTC m=+608.047442465" watchObservedRunningTime="2026-01-06 14:46:51.421065996 +0000 UTC m=+608.048532314" Jan 06 14:46:51 crc kubenswrapper[4744]: I0106 14:46:51.447202 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz" podStartSLOduration=17.625329175 podStartE2EDuration="23.447186543s" podCreationTimestamp="2026-01-06 14:46:28 +0000 UTC" firstStartedPulling="2026-01-06 14:46:44.980135133 +0000 UTC m=+601.607601451" lastFinishedPulling="2026-01-06 14:46:50.801992501 +0000 UTC m=+607.429458819" observedRunningTime="2026-01-06 14:46:51.444318967 +0000 UTC m=+608.071785285" watchObservedRunningTime="2026-01-06 14:46:51.447186543 +0000 UTC m=+608.074652861" Jan 06 14:46:52 crc kubenswrapper[4744]: I0106 14:46:52.349602 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" event={"ID":"7f4652da-367f-462e-92cb-cd4a7acc9487","Type":"ContainerStarted","Data":"ec56f0c627fd80df7277c1f580eb5762233c823011d9c22db8a225da4e7097c0"} Jan 06 14:46:55 crc kubenswrapper[4744]: I0106 14:46:55.366019 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" event={"ID":"7f4652da-367f-462e-92cb-cd4a7acc9487","Type":"ContainerStarted","Data":"4ce39de2434022e53f4d66405765c5a059ee75c723d22dd8dd243e3bcd99308e"} Jan 06 14:46:55 crc kubenswrapper[4744]: I0106 14:46:55.382449 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-q677r" podStartSLOduration=24.539156361 podStartE2EDuration="27.382405644s" podCreationTimestamp="2026-01-06 14:46:28 +0000 UTC" firstStartedPulling="2026-01-06 14:46:51.406190189 +0000 UTC m=+608.033656517" lastFinishedPulling="2026-01-06 14:46:54.249439482 +0000 UTC m=+610.876905800" observedRunningTime="2026-01-06 14:46:55.380891903 +0000 UTC m=+612.008358221" watchObservedRunningTime="2026-01-06 14:46:55.382405644 +0000 UTC m=+612.009871962" Jan 06 14:46:56 crc kubenswrapper[4744]: I0106 14:46:56.620232 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bffz5" Jan 06 14:46:57 crc kubenswrapper[4744]: I0106 14:46:57.389632 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" event={"ID":"313a5ded-3973-4f75-a2ff-d112d168de8d","Type":"ContainerStarted","Data":"ce2d78cec11cfec3c5d72d471d4699dc450bd8f7995fb8a02f99d5946d67cc7b"} Jan 06 14:46:57 crc kubenswrapper[4744]: I0106 14:46:57.389997 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:57 crc kubenswrapper[4744]: I0106 14:46:57.392238 4744 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-cklnq container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.7:8081/healthz\": dial tcp 10.217.0.7:8081: connect: connection refused" start-of-body= Jan 06 14:46:57 crc kubenswrapper[4744]: I0106 14:46:57.392322 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" podUID="313a5ded-3973-4f75-a2ff-d112d168de8d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.7:8081/healthz\": dial tcp 10.217.0.7:8081: connect: connection refused" Jan 06 14:46:58 crc kubenswrapper[4744]: I0106 14:46:58.453372 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" Jan 06 14:46:58 crc kubenswrapper[4744]: I0106 14:46:58.490886 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-cklnq" podStartSLOduration=24.670379374 podStartE2EDuration="30.490862856s" podCreationTimestamp="2026-01-06 14:46:28 +0000 UTC" firstStartedPulling="2026-01-06 14:46:51.323011869 +0000 UTC m=+607.950478187" lastFinishedPulling="2026-01-06 14:46:57.143495311 +0000 UTC m=+613.770961669" observedRunningTime="2026-01-06 14:46:57.416783996 +0000 UTC m=+614.044250314" watchObservedRunningTime="2026-01-06 14:46:58.490862856 +0000 UTC m=+615.118329214" Jan 06 14:46:58 crc kubenswrapper[4744]: I0106 14:46:58.918932 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-rx8pm" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.719325 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-66g5r"] Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.721182 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-66g5r" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.723330 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-nprtp" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.728713 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv8cq\" (UniqueName: \"kubernetes.io/projected/4947ffd9-d5ba-476d-89b8-c60b573f436d-kube-api-access-nv8cq\") pod \"cert-manager-cainjector-cf98fcc89-66g5r\" (UID: \"4947ffd9-d5ba-476d-89b8-c60b573f436d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-66g5r" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.733143 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.733755 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.734432 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-9gjwj"] Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.735471 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-9gjwj" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.739468 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-nrsjx" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.747274 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-66g5r"] Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.755313 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-tvhp5"] Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.756366 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.760570 4744 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-g6sbp" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.770250 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-9gjwj"] Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.792131 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-tvhp5"] Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.829316 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5md6d\" (UniqueName: \"kubernetes.io/projected/9539eba1-94db-41e9-80ed-ba1de1978f94-kube-api-access-5md6d\") pod \"cert-manager-webhook-687f57d79b-tvhp5\" (UID: \"9539eba1-94db-41e9-80ed-ba1de1978f94\") " pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.829598 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv8cq\" (UniqueName: \"kubernetes.io/projected/4947ffd9-d5ba-476d-89b8-c60b573f436d-kube-api-access-nv8cq\") pod \"cert-manager-cainjector-cf98fcc89-66g5r\" (UID: \"4947ffd9-d5ba-476d-89b8-c60b573f436d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-66g5r" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.829697 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4jlx\" (UniqueName: \"kubernetes.io/projected/c472e993-f700-40e9-9a30-6d81b169cd5c-kube-api-access-f4jlx\") pod \"cert-manager-858654f9db-9gjwj\" (UID: \"c472e993-f700-40e9-9a30-6d81b169cd5c\") " pod="cert-manager/cert-manager-858654f9db-9gjwj" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.859553 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv8cq\" (UniqueName: \"kubernetes.io/projected/4947ffd9-d5ba-476d-89b8-c60b573f436d-kube-api-access-nv8cq\") pod \"cert-manager-cainjector-cf98fcc89-66g5r\" (UID: \"4947ffd9-d5ba-476d-89b8-c60b573f436d\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-66g5r" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.930676 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5md6d\" (UniqueName: \"kubernetes.io/projected/9539eba1-94db-41e9-80ed-ba1de1978f94-kube-api-access-5md6d\") pod \"cert-manager-webhook-687f57d79b-tvhp5\" (UID: \"9539eba1-94db-41e9-80ed-ba1de1978f94\") " pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.930729 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4jlx\" (UniqueName: \"kubernetes.io/projected/c472e993-f700-40e9-9a30-6d81b169cd5c-kube-api-access-f4jlx\") pod \"cert-manager-858654f9db-9gjwj\" (UID: \"c472e993-f700-40e9-9a30-6d81b169cd5c\") " pod="cert-manager/cert-manager-858654f9db-9gjwj" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.947514 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4jlx\" (UniqueName: \"kubernetes.io/projected/c472e993-f700-40e9-9a30-6d81b169cd5c-kube-api-access-f4jlx\") pod \"cert-manager-858654f9db-9gjwj\" (UID: \"c472e993-f700-40e9-9a30-6d81b169cd5c\") " pod="cert-manager/cert-manager-858654f9db-9gjwj" Jan 06 14:47:07 crc kubenswrapper[4744]: I0106 14:47:07.953878 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5md6d\" (UniqueName: \"kubernetes.io/projected/9539eba1-94db-41e9-80ed-ba1de1978f94-kube-api-access-5md6d\") pod \"cert-manager-webhook-687f57d79b-tvhp5\" (UID: \"9539eba1-94db-41e9-80ed-ba1de1978f94\") " pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" Jan 06 14:47:08 crc kubenswrapper[4744]: I0106 14:47:08.044082 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-66g5r" Jan 06 14:47:08 crc kubenswrapper[4744]: I0106 14:47:08.062125 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-9gjwj" Jan 06 14:47:08 crc kubenswrapper[4744]: I0106 14:47:08.075947 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" Jan 06 14:47:08 crc kubenswrapper[4744]: I0106 14:47:08.556557 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-66g5r"] Jan 06 14:47:08 crc kubenswrapper[4744]: W0106 14:47:08.560494 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4947ffd9_d5ba_476d_89b8_c60b573f436d.slice/crio-25faad42192cd42f49d32e89cb6f63a7c5d58f17f85970c51799763ae23ec3b6 WatchSource:0}: Error finding container 25faad42192cd42f49d32e89cb6f63a7c5d58f17f85970c51799763ae23ec3b6: Status 404 returned error can't find the container with id 25faad42192cd42f49d32e89cb6f63a7c5d58f17f85970c51799763ae23ec3b6 Jan 06 14:47:08 crc kubenswrapper[4744]: I0106 14:47:08.603830 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-9gjwj"] Jan 06 14:47:08 crc kubenswrapper[4744]: I0106 14:47:08.619021 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-tvhp5"] Jan 06 14:47:08 crc kubenswrapper[4744]: W0106 14:47:08.621269 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9539eba1_94db_41e9_80ed_ba1de1978f94.slice/crio-f0b77a4f2eb00ecc828880fb431fadaa838a6a12a60c2769bc2baed19f1e7e64 WatchSource:0}: Error finding container f0b77a4f2eb00ecc828880fb431fadaa838a6a12a60c2769bc2baed19f1e7e64: Status 404 returned error can't find the container with id f0b77a4f2eb00ecc828880fb431fadaa838a6a12a60c2769bc2baed19f1e7e64 Jan 06 14:47:09 crc kubenswrapper[4744]: I0106 14:47:09.484969 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-9gjwj" event={"ID":"c472e993-f700-40e9-9a30-6d81b169cd5c","Type":"ContainerStarted","Data":"f24401829e11305245a1af3368c1cbd488f46862889116d6a74529cc20a0f8e8"} Jan 06 14:47:09 crc kubenswrapper[4744]: I0106 14:47:09.487701 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" event={"ID":"9539eba1-94db-41e9-80ed-ba1de1978f94","Type":"ContainerStarted","Data":"f0b77a4f2eb00ecc828880fb431fadaa838a6a12a60c2769bc2baed19f1e7e64"} Jan 06 14:47:09 crc kubenswrapper[4744]: I0106 14:47:09.490026 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-66g5r" event={"ID":"4947ffd9-d5ba-476d-89b8-c60b573f436d","Type":"ContainerStarted","Data":"25faad42192cd42f49d32e89cb6f63a7c5d58f17f85970c51799763ae23ec3b6"} Jan 06 14:47:13 crc kubenswrapper[4744]: I0106 14:47:13.522050 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" event={"ID":"9539eba1-94db-41e9-80ed-ba1de1978f94","Type":"ContainerStarted","Data":"a3420215fa51a8fd9156cd16662649903cae5a752452a59486db96afa518fcd9"} Jan 06 14:47:13 crc kubenswrapper[4744]: I0106 14:47:13.522586 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" Jan 06 14:47:13 crc kubenswrapper[4744]: I0106 14:47:13.524131 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-66g5r" event={"ID":"4947ffd9-d5ba-476d-89b8-c60b573f436d","Type":"ContainerStarted","Data":"460879becab8c176f8d3e842e277a112b924e4ce5b2dd4db94de8832ce7547f7"} Jan 06 14:47:13 crc kubenswrapper[4744]: I0106 14:47:13.525779 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-9gjwj" event={"ID":"c472e993-f700-40e9-9a30-6d81b169cd5c","Type":"ContainerStarted","Data":"8ff15850f20ebd3edd14349922704f921e1ec25b9e46bf1097109bafebce70dd"} Jan 06 14:47:13 crc kubenswrapper[4744]: I0106 14:47:13.549885 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" podStartSLOduration=1.9698607849999998 podStartE2EDuration="6.549868096s" podCreationTimestamp="2026-01-06 14:47:07 +0000 UTC" firstStartedPulling="2026-01-06 14:47:08.623629963 +0000 UTC m=+625.251096281" lastFinishedPulling="2026-01-06 14:47:13.203637234 +0000 UTC m=+629.831103592" observedRunningTime="2026-01-06 14:47:13.549069545 +0000 UTC m=+630.176535863" watchObservedRunningTime="2026-01-06 14:47:13.549868096 +0000 UTC m=+630.177334424" Jan 06 14:47:13 crc kubenswrapper[4744]: I0106 14:47:13.588540 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-9gjwj" podStartSLOduration=2.014971739 podStartE2EDuration="6.588505217s" podCreationTimestamp="2026-01-06 14:47:07 +0000 UTC" firstStartedPulling="2026-01-06 14:47:08.604107422 +0000 UTC m=+625.231573790" lastFinishedPulling="2026-01-06 14:47:13.17764091 +0000 UTC m=+629.805107268" observedRunningTime="2026-01-06 14:47:13.572635654 +0000 UTC m=+630.200102022" watchObservedRunningTime="2026-01-06 14:47:13.588505217 +0000 UTC m=+630.215971575" Jan 06 14:47:18 crc kubenswrapper[4744]: I0106 14:47:18.116600 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-tvhp5" Jan 06 14:47:18 crc kubenswrapper[4744]: I0106 14:47:18.148822 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-66g5r" podStartSLOduration=6.573382752 podStartE2EDuration="11.148791841s" podCreationTimestamp="2026-01-06 14:47:07 +0000 UTC" firstStartedPulling="2026-01-06 14:47:08.563037745 +0000 UTC m=+625.190504103" lastFinishedPulling="2026-01-06 14:47:13.138446834 +0000 UTC m=+629.765913192" observedRunningTime="2026-01-06 14:47:13.595845153 +0000 UTC m=+630.223311481" watchObservedRunningTime="2026-01-06 14:47:18.148791841 +0000 UTC m=+634.776258169" Jan 06 14:47:29 crc kubenswrapper[4744]: I0106 14:47:29.966228 4744 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 06 14:47:42 crc kubenswrapper[4744]: I0106 14:47:42.954700 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6"] Jan 06 14:47:42 crc kubenswrapper[4744]: I0106 14:47:42.957371 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:42 crc kubenswrapper[4744]: I0106 14:47:42.962046 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 06 14:47:42 crc kubenswrapper[4744]: I0106 14:47:42.974112 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6"] Jan 06 14:47:42 crc kubenswrapper[4744]: I0106 14:47:42.975123 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-util\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:42 crc kubenswrapper[4744]: I0106 14:47:42.975254 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-bundle\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:42 crc kubenswrapper[4744]: I0106 14:47:42.975343 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nb9b\" (UniqueName: \"kubernetes.io/projected/69e01720-c541-4435-8ea7-3012016b80ff-kube-api-access-2nb9b\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.076221 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-bundle\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.076286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nb9b\" (UniqueName: \"kubernetes.io/projected/69e01720-c541-4435-8ea7-3012016b80ff-kube-api-access-2nb9b\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.076336 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-util\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.076791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-util\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.077552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-bundle\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.098478 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nb9b\" (UniqueName: \"kubernetes.io/projected/69e01720-c541-4435-8ea7-3012016b80ff-kube-api-access-2nb9b\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.293034 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.348805 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2"] Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.350152 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.359256 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2"] Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.379692 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-bundle\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.379766 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-util\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.379812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2msrv\" (UniqueName: \"kubernetes.io/projected/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-kube-api-access-2msrv\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.481512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-bundle\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.481641 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-util\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.481690 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2msrv\" (UniqueName: \"kubernetes.io/projected/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-kube-api-access-2msrv\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.482400 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-util\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.483599 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-bundle\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.503840 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2msrv\" (UniqueName: \"kubernetes.io/projected/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-kube-api-access-2msrv\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.674509 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:43 crc kubenswrapper[4744]: I0106 14:47:43.747043 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6"] Jan 06 14:47:43 crc kubenswrapper[4744]: W0106 14:47:43.759876 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69e01720_c541_4435_8ea7_3012016b80ff.slice/crio-7f2f35cbbb3b43784ad9e1c1d0d6eb38089f14ee4c61ac8b97a49cb75831d03a WatchSource:0}: Error finding container 7f2f35cbbb3b43784ad9e1c1d0d6eb38089f14ee4c61ac8b97a49cb75831d03a: Status 404 returned error can't find the container with id 7f2f35cbbb3b43784ad9e1c1d0d6eb38089f14ee4c61ac8b97a49cb75831d03a Jan 06 14:47:44 crc kubenswrapper[4744]: I0106 14:47:44.170821 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2"] Jan 06 14:47:44 crc kubenswrapper[4744]: W0106 14:47:44.191027 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fc7c9b1_f812_4104_8a87_b9d3024fc25e.slice/crio-cfcff79eb962d1755afa5e82354a12e24f717b8c0e1755ab0929cfc6f96b65f9 WatchSource:0}: Error finding container cfcff79eb962d1755afa5e82354a12e24f717b8c0e1755ab0929cfc6f96b65f9: Status 404 returned error can't find the container with id cfcff79eb962d1755afa5e82354a12e24f717b8c0e1755ab0929cfc6f96b65f9 Jan 06 14:47:44 crc kubenswrapper[4744]: I0106 14:47:44.775230 4744 generic.go:334] "Generic (PLEG): container finished" podID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerID="1a13933501da9b17071f9b2cb9d80192d9459c9c24b3e994d31b3fc50b22a080" exitCode=0 Jan 06 14:47:44 crc kubenswrapper[4744]: I0106 14:47:44.775303 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" event={"ID":"9fc7c9b1-f812-4104-8a87-b9d3024fc25e","Type":"ContainerDied","Data":"1a13933501da9b17071f9b2cb9d80192d9459c9c24b3e994d31b3fc50b22a080"} Jan 06 14:47:44 crc kubenswrapper[4744]: I0106 14:47:44.775341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" event={"ID":"9fc7c9b1-f812-4104-8a87-b9d3024fc25e","Type":"ContainerStarted","Data":"cfcff79eb962d1755afa5e82354a12e24f717b8c0e1755ab0929cfc6f96b65f9"} Jan 06 14:47:44 crc kubenswrapper[4744]: I0106 14:47:44.778545 4744 generic.go:334] "Generic (PLEG): container finished" podID="69e01720-c541-4435-8ea7-3012016b80ff" containerID="a8325872f29bc71ad7db1e07039adb73e8a5965274971ae735e435c79026f55d" exitCode=0 Jan 06 14:47:44 crc kubenswrapper[4744]: I0106 14:47:44.778583 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" event={"ID":"69e01720-c541-4435-8ea7-3012016b80ff","Type":"ContainerDied","Data":"a8325872f29bc71ad7db1e07039adb73e8a5965274971ae735e435c79026f55d"} Jan 06 14:47:44 crc kubenswrapper[4744]: I0106 14:47:44.778609 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" event={"ID":"69e01720-c541-4435-8ea7-3012016b80ff","Type":"ContainerStarted","Data":"7f2f35cbbb3b43784ad9e1c1d0d6eb38089f14ee4c61ac8b97a49cb75831d03a"} Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.700578 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-75v4p"] Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.703782 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.729356 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-75v4p"] Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.764791 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-utilities\") pod \"redhat-operators-75v4p\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.764906 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-catalog-content\") pod \"redhat-operators-75v4p\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.764960 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpg7q\" (UniqueName: \"kubernetes.io/projected/1a20e146-901c-4666-8886-c6e26ce8b521-kube-api-access-cpg7q\") pod \"redhat-operators-75v4p\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.866815 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-catalog-content\") pod \"redhat-operators-75v4p\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.866908 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpg7q\" (UniqueName: \"kubernetes.io/projected/1a20e146-901c-4666-8886-c6e26ce8b521-kube-api-access-cpg7q\") pod \"redhat-operators-75v4p\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.866964 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-utilities\") pod \"redhat-operators-75v4p\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.867625 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-utilities\") pod \"redhat-operators-75v4p\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.867834 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-catalog-content\") pod \"redhat-operators-75v4p\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:46 crc kubenswrapper[4744]: I0106 14:47:46.891926 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpg7q\" (UniqueName: \"kubernetes.io/projected/1a20e146-901c-4666-8886-c6e26ce8b521-kube-api-access-cpg7q\") pod \"redhat-operators-75v4p\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:47 crc kubenswrapper[4744]: I0106 14:47:47.066097 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:47 crc kubenswrapper[4744]: I0106 14:47:47.382176 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-75v4p"] Jan 06 14:47:47 crc kubenswrapper[4744]: W0106 14:47:47.393101 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a20e146_901c_4666_8886_c6e26ce8b521.slice/crio-b40664c0703ca2ee73b358579b14c732c5788ea8256c434d56368d01ca5dd042 WatchSource:0}: Error finding container b40664c0703ca2ee73b358579b14c732c5788ea8256c434d56368d01ca5dd042: Status 404 returned error can't find the container with id b40664c0703ca2ee73b358579b14c732c5788ea8256c434d56368d01ca5dd042 Jan 06 14:47:47 crc kubenswrapper[4744]: I0106 14:47:47.802682 4744 generic.go:334] "Generic (PLEG): container finished" podID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerID="618bdc8dde5df59602a57819494e8e068c9297d857c37e84c667e94b0148ccbd" exitCode=0 Jan 06 14:47:47 crc kubenswrapper[4744]: I0106 14:47:47.802744 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" event={"ID":"9fc7c9b1-f812-4104-8a87-b9d3024fc25e","Type":"ContainerDied","Data":"618bdc8dde5df59602a57819494e8e068c9297d857c37e84c667e94b0148ccbd"} Jan 06 14:47:47 crc kubenswrapper[4744]: I0106 14:47:47.803625 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75v4p" event={"ID":"1a20e146-901c-4666-8886-c6e26ce8b521","Type":"ContainerStarted","Data":"b40664c0703ca2ee73b358579b14c732c5788ea8256c434d56368d01ca5dd042"} Jan 06 14:47:48 crc kubenswrapper[4744]: I0106 14:47:48.816229 4744 generic.go:334] "Generic (PLEG): container finished" podID="1a20e146-901c-4666-8886-c6e26ce8b521" containerID="6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016" exitCode=0 Jan 06 14:47:48 crc kubenswrapper[4744]: I0106 14:47:48.816359 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75v4p" event={"ID":"1a20e146-901c-4666-8886-c6e26ce8b521","Type":"ContainerDied","Data":"6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016"} Jan 06 14:47:49 crc kubenswrapper[4744]: I0106 14:47:49.825127 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75v4p" event={"ID":"1a20e146-901c-4666-8886-c6e26ce8b521","Type":"ContainerStarted","Data":"9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d"} Jan 06 14:47:49 crc kubenswrapper[4744]: I0106 14:47:49.827474 4744 generic.go:334] "Generic (PLEG): container finished" podID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerID="5c9c1409f03dd6f92efae92218a732f293e2f560874fd62b402fdab5458aea76" exitCode=0 Jan 06 14:47:49 crc kubenswrapper[4744]: I0106 14:47:49.827547 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" event={"ID":"9fc7c9b1-f812-4104-8a87-b9d3024fc25e","Type":"ContainerDied","Data":"5c9c1409f03dd6f92efae92218a732f293e2f560874fd62b402fdab5458aea76"} Jan 06 14:47:50 crc kubenswrapper[4744]: I0106 14:47:50.835646 4744 generic.go:334] "Generic (PLEG): container finished" podID="1a20e146-901c-4666-8886-c6e26ce8b521" containerID="9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d" exitCode=0 Jan 06 14:47:50 crc kubenswrapper[4744]: I0106 14:47:50.835747 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75v4p" event={"ID":"1a20e146-901c-4666-8886-c6e26ce8b521","Type":"ContainerDied","Data":"9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d"} Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.170107 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.234080 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-bundle\") pod \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.234278 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-util\") pod \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.234403 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2msrv\" (UniqueName: \"kubernetes.io/projected/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-kube-api-access-2msrv\") pod \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\" (UID: \"9fc7c9b1-f812-4104-8a87-b9d3024fc25e\") " Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.235692 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-bundle" (OuterVolumeSpecName: "bundle") pod "9fc7c9b1-f812-4104-8a87-b9d3024fc25e" (UID: "9fc7c9b1-f812-4104-8a87-b9d3024fc25e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.243851 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-kube-api-access-2msrv" (OuterVolumeSpecName: "kube-api-access-2msrv") pod "9fc7c9b1-f812-4104-8a87-b9d3024fc25e" (UID: "9fc7c9b1-f812-4104-8a87-b9d3024fc25e"). InnerVolumeSpecName "kube-api-access-2msrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.249321 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-util" (OuterVolumeSpecName: "util") pod "9fc7c9b1-f812-4104-8a87-b9d3024fc25e" (UID: "9fc7c9b1-f812-4104-8a87-b9d3024fc25e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.336718 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-util\") on node \"crc\" DevicePath \"\"" Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.336771 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2msrv\" (UniqueName: \"kubernetes.io/projected/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-kube-api-access-2msrv\") on node \"crc\" DevicePath \"\"" Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.336794 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c9b1-f812-4104-8a87-b9d3024fc25e-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.848410 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" event={"ID":"9fc7c9b1-f812-4104-8a87-b9d3024fc25e","Type":"ContainerDied","Data":"cfcff79eb962d1755afa5e82354a12e24f717b8c0e1755ab0929cfc6f96b65f9"} Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.848715 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfcff79eb962d1755afa5e82354a12e24f717b8c0e1755ab0929cfc6f96b65f9" Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.848496 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2" Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.852790 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75v4p" event={"ID":"1a20e146-901c-4666-8886-c6e26ce8b521","Type":"ContainerStarted","Data":"1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32"} Jan 06 14:47:51 crc kubenswrapper[4744]: I0106 14:47:51.875664 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-75v4p" podStartSLOduration=3.388305775 podStartE2EDuration="5.875639513s" podCreationTimestamp="2026-01-06 14:47:46 +0000 UTC" firstStartedPulling="2026-01-06 14:47:48.820185779 +0000 UTC m=+665.447652147" lastFinishedPulling="2026-01-06 14:47:51.307519557 +0000 UTC m=+667.934985885" observedRunningTime="2026-01-06 14:47:51.872835019 +0000 UTC m=+668.500301377" watchObservedRunningTime="2026-01-06 14:47:51.875639513 +0000 UTC m=+668.503105871" Jan 06 14:47:57 crc kubenswrapper[4744]: I0106 14:47:57.067506 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:57 crc kubenswrapper[4744]: I0106 14:47:57.068317 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.148334 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-75v4p" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" containerName="registry-server" probeResult="failure" output=< Jan 06 14:47:58 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:47:58 crc kubenswrapper[4744]: > Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.978654 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb"] Jan 06 14:47:58 crc kubenswrapper[4744]: E0106 14:47:58.978972 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerName="util" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.978994 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerName="util" Jan 06 14:47:58 crc kubenswrapper[4744]: E0106 14:47:58.979013 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerName="pull" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.979022 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerName="pull" Jan 06 14:47:58 crc kubenswrapper[4744]: E0106 14:47:58.979045 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerName="extract" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.979054 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerName="extract" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.979237 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fc7c9b1-f812-4104-8a87-b9d3024fc25e" containerName="extract" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.980092 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.988002 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.988049 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.988129 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.988345 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-c7wdh" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.988386 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.990364 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Jan 06 14:47:58 crc kubenswrapper[4744]: I0106 14:47:58.998197 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb"] Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.057565 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/34926775-1ace-452c-ae93-f25253e4b811-webhook-cert\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.057634 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34926775-1ace-452c-ae93-f25253e4b811-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.057751 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llt6b\" (UniqueName: \"kubernetes.io/projected/34926775-1ace-452c-ae93-f25253e4b811-kube-api-access-llt6b\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.058018 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/34926775-1ace-452c-ae93-f25253e4b811-manager-config\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.058240 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/34926775-1ace-452c-ae93-f25253e4b811-apiservice-cert\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.160887 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/34926775-1ace-452c-ae93-f25253e4b811-webhook-cert\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.160972 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34926775-1ace-452c-ae93-f25253e4b811-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.161008 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llt6b\" (UniqueName: \"kubernetes.io/projected/34926775-1ace-452c-ae93-f25253e4b811-kube-api-access-llt6b\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.161248 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/34926775-1ace-452c-ae93-f25253e4b811-manager-config\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.161298 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/34926775-1ace-452c-ae93-f25253e4b811-apiservice-cert\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.164790 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/34926775-1ace-452c-ae93-f25253e4b811-manager-config\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.168576 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/34926775-1ace-452c-ae93-f25253e4b811-apiservice-cert\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.169827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/34926775-1ace-452c-ae93-f25253e4b811-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.169856 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/34926775-1ace-452c-ae93-f25253e4b811-webhook-cert\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.188518 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llt6b\" (UniqueName: \"kubernetes.io/projected/34926775-1ace-452c-ae93-f25253e4b811-kube-api-access-llt6b\") pod \"loki-operator-controller-manager-55d9dcdcb9-txlgb\" (UID: \"34926775-1ace-452c-ae93-f25253e4b811\") " pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.298882 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.730420 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb"] Jan 06 14:47:59 crc kubenswrapper[4744]: I0106 14:47:59.911149 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" event={"ID":"34926775-1ace-452c-ae93-f25253e4b811","Type":"ContainerStarted","Data":"b25e4e64a637708a94725f02f5c0f116b6c03bc7bd221f89631e32f86efecb42"} Jan 06 14:48:05 crc kubenswrapper[4744]: I0106 14:48:05.956243 4744 generic.go:334] "Generic (PLEG): container finished" podID="69e01720-c541-4435-8ea7-3012016b80ff" containerID="753d4d00b1253526eacb368b92b6c3e8674c841e4bb21cb32758582880c43dfb" exitCode=0 Jan 06 14:48:05 crc kubenswrapper[4744]: I0106 14:48:05.956435 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" event={"ID":"69e01720-c541-4435-8ea7-3012016b80ff","Type":"ContainerDied","Data":"753d4d00b1253526eacb368b92b6c3e8674c841e4bb21cb32758582880c43dfb"} Jan 06 14:48:06 crc kubenswrapper[4744]: I0106 14:48:06.991796 4744 generic.go:334] "Generic (PLEG): container finished" podID="69e01720-c541-4435-8ea7-3012016b80ff" containerID="8505329f62412ddfb92aae2a668b7a5b246deddc883806af3991d9221c5663d2" exitCode=0 Jan 06 14:48:06 crc kubenswrapper[4744]: I0106 14:48:06.992035 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" event={"ID":"69e01720-c541-4435-8ea7-3012016b80ff","Type":"ContainerDied","Data":"8505329f62412ddfb92aae2a668b7a5b246deddc883806af3991d9221c5663d2"} Jan 06 14:48:07 crc kubenswrapper[4744]: I0106 14:48:07.109660 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:48:07 crc kubenswrapper[4744]: I0106 14:48:07.154280 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:48:08 crc kubenswrapper[4744]: I0106 14:48:08.486667 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-75v4p"] Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.009097 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" event={"ID":"69e01720-c541-4435-8ea7-3012016b80ff","Type":"ContainerDied","Data":"7f2f35cbbb3b43784ad9e1c1d0d6eb38089f14ee4c61ac8b97a49cb75831d03a"} Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.009142 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f2f35cbbb3b43784ad9e1c1d0d6eb38089f14ee4c61ac8b97a49cb75831d03a" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.009487 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-75v4p" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" containerName="registry-server" containerID="cri-o://1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32" gracePeriod=2 Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.028856 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.113310 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-bundle\") pod \"69e01720-c541-4435-8ea7-3012016b80ff\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.113656 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nb9b\" (UniqueName: \"kubernetes.io/projected/69e01720-c541-4435-8ea7-3012016b80ff-kube-api-access-2nb9b\") pod \"69e01720-c541-4435-8ea7-3012016b80ff\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.113759 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-util\") pod \"69e01720-c541-4435-8ea7-3012016b80ff\" (UID: \"69e01720-c541-4435-8ea7-3012016b80ff\") " Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.118378 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-bundle" (OuterVolumeSpecName: "bundle") pod "69e01720-c541-4435-8ea7-3012016b80ff" (UID: "69e01720-c541-4435-8ea7-3012016b80ff"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.130773 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69e01720-c541-4435-8ea7-3012016b80ff-kube-api-access-2nb9b" (OuterVolumeSpecName: "kube-api-access-2nb9b") pod "69e01720-c541-4435-8ea7-3012016b80ff" (UID: "69e01720-c541-4435-8ea7-3012016b80ff"). InnerVolumeSpecName "kube-api-access-2nb9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.136258 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-util" (OuterVolumeSpecName: "util") pod "69e01720-c541-4435-8ea7-3012016b80ff" (UID: "69e01720-c541-4435-8ea7-3012016b80ff"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.216886 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-util\") on node \"crc\" DevicePath \"\"" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.216924 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/69e01720-c541-4435-8ea7-3012016b80ff-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.216934 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nb9b\" (UniqueName: \"kubernetes.io/projected/69e01720-c541-4435-8ea7-3012016b80ff-kube-api-access-2nb9b\") on node \"crc\" DevicePath \"\"" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.465450 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.520478 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-catalog-content\") pod \"1a20e146-901c-4666-8886-c6e26ce8b521\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.520549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-utilities\") pod \"1a20e146-901c-4666-8886-c6e26ce8b521\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.520685 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpg7q\" (UniqueName: \"kubernetes.io/projected/1a20e146-901c-4666-8886-c6e26ce8b521-kube-api-access-cpg7q\") pod \"1a20e146-901c-4666-8886-c6e26ce8b521\" (UID: \"1a20e146-901c-4666-8886-c6e26ce8b521\") " Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.522813 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-utilities" (OuterVolumeSpecName: "utilities") pod "1a20e146-901c-4666-8886-c6e26ce8b521" (UID: "1a20e146-901c-4666-8886-c6e26ce8b521"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.527307 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a20e146-901c-4666-8886-c6e26ce8b521-kube-api-access-cpg7q" (OuterVolumeSpecName: "kube-api-access-cpg7q") pod "1a20e146-901c-4666-8886-c6e26ce8b521" (UID: "1a20e146-901c-4666-8886-c6e26ce8b521"). InnerVolumeSpecName "kube-api-access-cpg7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.623943 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.623972 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpg7q\" (UniqueName: \"kubernetes.io/projected/1a20e146-901c-4666-8886-c6e26ce8b521-kube-api-access-cpg7q\") on node \"crc\" DevicePath \"\"" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.645681 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a20e146-901c-4666-8886-c6e26ce8b521" (UID: "1a20e146-901c-4666-8886-c6e26ce8b521"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:48:09 crc kubenswrapper[4744]: I0106 14:48:09.724875 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a20e146-901c-4666-8886-c6e26ce8b521-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.021024 4744 generic.go:334] "Generic (PLEG): container finished" podID="1a20e146-901c-4666-8886-c6e26ce8b521" containerID="1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32" exitCode=0 Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.021379 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-75v4p" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.021382 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75v4p" event={"ID":"1a20e146-901c-4666-8886-c6e26ce8b521","Type":"ContainerDied","Data":"1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32"} Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.021708 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-75v4p" event={"ID":"1a20e146-901c-4666-8886-c6e26ce8b521","Type":"ContainerDied","Data":"b40664c0703ca2ee73b358579b14c732c5788ea8256c434d56368d01ca5dd042"} Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.021857 4744 scope.go:117] "RemoveContainer" containerID="1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.024105 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.024780 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" event={"ID":"34926775-1ace-452c-ae93-f25253e4b811","Type":"ContainerStarted","Data":"068728c8fb0d61b6bbd65e9da3f9a13ffe91c5527a72350ca6995bbe94399217"} Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.060190 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-75v4p"] Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.073299 4744 scope.go:117] "RemoveContainer" containerID="9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.081790 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-75v4p"] Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.106685 4744 scope.go:117] "RemoveContainer" containerID="6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.134565 4744 scope.go:117] "RemoveContainer" containerID="1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32" Jan 06 14:48:10 crc kubenswrapper[4744]: E0106 14:48:10.135116 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32\": container with ID starting with 1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32 not found: ID does not exist" containerID="1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.135179 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32"} err="failed to get container status \"1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32\": rpc error: code = NotFound desc = could not find container \"1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32\": container with ID starting with 1baf38109e41ac1d6dda1eaa4ce05c7c3f900f5be76ebe3aa18ebddd7761de32 not found: ID does not exist" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.135213 4744 scope.go:117] "RemoveContainer" containerID="9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d" Jan 06 14:48:10 crc kubenswrapper[4744]: E0106 14:48:10.136068 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d\": container with ID starting with 9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d not found: ID does not exist" containerID="9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.136109 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d"} err="failed to get container status \"9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d\": rpc error: code = NotFound desc = could not find container \"9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d\": container with ID starting with 9d6059e583cda6fd04807a6484ef6485a21c09d50d85cf23dd7e824abcf2757d not found: ID does not exist" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.136128 4744 scope.go:117] "RemoveContainer" containerID="6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016" Jan 06 14:48:10 crc kubenswrapper[4744]: E0106 14:48:10.136481 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016\": container with ID starting with 6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016 not found: ID does not exist" containerID="6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016" Jan 06 14:48:10 crc kubenswrapper[4744]: I0106 14:48:10.136556 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016"} err="failed to get container status \"6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016\": rpc error: code = NotFound desc = could not find container \"6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016\": container with ID starting with 6b4a69127e8d3939598f6bb59bc7eaeb9d7f10ae4d12279bb0e0c1cce4fcd016 not found: ID does not exist" Jan 06 14:48:11 crc kubenswrapper[4744]: I0106 14:48:11.723810 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" path="/var/lib/kubelet/pods/1a20e146-901c-4666-8886-c6e26ce8b521/volumes" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.423964 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.424270 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.670243 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck"] Jan 06 14:48:14 crc kubenswrapper[4744]: E0106 14:48:14.670523 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e01720-c541-4435-8ea7-3012016b80ff" containerName="extract" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.670538 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e01720-c541-4435-8ea7-3012016b80ff" containerName="extract" Jan 06 14:48:14 crc kubenswrapper[4744]: E0106 14:48:14.670560 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e01720-c541-4435-8ea7-3012016b80ff" containerName="util" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.670569 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e01720-c541-4435-8ea7-3012016b80ff" containerName="util" Jan 06 14:48:14 crc kubenswrapper[4744]: E0106 14:48:14.670580 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" containerName="extract-content" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.670588 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" containerName="extract-content" Jan 06 14:48:14 crc kubenswrapper[4744]: E0106 14:48:14.670609 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" containerName="extract-utilities" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.670618 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" containerName="extract-utilities" Jan 06 14:48:14 crc kubenswrapper[4744]: E0106 14:48:14.670634 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" containerName="registry-server" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.670642 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" containerName="registry-server" Jan 06 14:48:14 crc kubenswrapper[4744]: E0106 14:48:14.670655 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e01720-c541-4435-8ea7-3012016b80ff" containerName="pull" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.670664 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e01720-c541-4435-8ea7-3012016b80ff" containerName="pull" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.670836 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="69e01720-c541-4435-8ea7-3012016b80ff" containerName="extract" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.670857 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a20e146-901c-4666-8886-c6e26ce8b521" containerName="registry-server" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.671442 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.674521 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.674620 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-898t6" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.686375 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.686761 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck"] Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.703555 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fw8t\" (UniqueName: \"kubernetes.io/projected/8a59d77a-d5d2-4a5c-a63d-cdacacf67f0b-kube-api-access-7fw8t\") pod \"cluster-logging-operator-79cf69ddc8-4qxck\" (UID: \"8a59d77a-d5d2-4a5c-a63d-cdacacf67f0b\") " pod="openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.805766 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fw8t\" (UniqueName: \"kubernetes.io/projected/8a59d77a-d5d2-4a5c-a63d-cdacacf67f0b-kube-api-access-7fw8t\") pod \"cluster-logging-operator-79cf69ddc8-4qxck\" (UID: \"8a59d77a-d5d2-4a5c-a63d-cdacacf67f0b\") " pod="openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck" Jan 06 14:48:14 crc kubenswrapper[4744]: I0106 14:48:14.827067 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fw8t\" (UniqueName: \"kubernetes.io/projected/8a59d77a-d5d2-4a5c-a63d-cdacacf67f0b-kube-api-access-7fw8t\") pod \"cluster-logging-operator-79cf69ddc8-4qxck\" (UID: \"8a59d77a-d5d2-4a5c-a63d-cdacacf67f0b\") " pod="openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck" Jan 06 14:48:15 crc kubenswrapper[4744]: I0106 14:48:15.040351 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck" Jan 06 14:48:15 crc kubenswrapper[4744]: I0106 14:48:15.775392 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck"] Jan 06 14:48:15 crc kubenswrapper[4744]: W0106 14:48:15.784609 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a59d77a_d5d2_4a5c_a63d_cdacacf67f0b.slice/crio-45b34438cedf2ae2bb3cf076391b6e853248b4b67c164893658b2716e789f6f6 WatchSource:0}: Error finding container 45b34438cedf2ae2bb3cf076391b6e853248b4b67c164893658b2716e789f6f6: Status 404 returned error can't find the container with id 45b34438cedf2ae2bb3cf076391b6e853248b4b67c164893658b2716e789f6f6 Jan 06 14:48:16 crc kubenswrapper[4744]: I0106 14:48:16.067574 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" event={"ID":"34926775-1ace-452c-ae93-f25253e4b811","Type":"ContainerStarted","Data":"f563caf99c43a4fa15f6f775312f844cd207e6cc9cd720796fcfdb36e66a5e15"} Jan 06 14:48:16 crc kubenswrapper[4744]: I0106 14:48:16.068089 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:48:16 crc kubenswrapper[4744]: I0106 14:48:16.069035 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck" event={"ID":"8a59d77a-d5d2-4a5c-a63d-cdacacf67f0b","Type":"ContainerStarted","Data":"45b34438cedf2ae2bb3cf076391b6e853248b4b67c164893658b2716e789f6f6"} Jan 06 14:48:16 crc kubenswrapper[4744]: I0106 14:48:16.071254 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" Jan 06 14:48:16 crc kubenswrapper[4744]: I0106 14:48:16.104025 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-55d9dcdcb9-txlgb" podStartSLOduration=2.234680433 podStartE2EDuration="18.104000322s" podCreationTimestamp="2026-01-06 14:47:58 +0000 UTC" firstStartedPulling="2026-01-06 14:47:59.733729921 +0000 UTC m=+676.361196239" lastFinishedPulling="2026-01-06 14:48:15.60304981 +0000 UTC m=+692.230516128" observedRunningTime="2026-01-06 14:48:16.094938663 +0000 UTC m=+692.722405001" watchObservedRunningTime="2026-01-06 14:48:16.104000322 +0000 UTC m=+692.731466670" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.658949 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c6nkb"] Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.661706 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.676891 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c6nkb"] Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.774331 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6vrb\" (UniqueName: \"kubernetes.io/projected/35ec5155-43ef-4861-a925-28e3688546fe-kube-api-access-c6vrb\") pod \"certified-operators-c6nkb\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.774653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-utilities\") pod \"certified-operators-c6nkb\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.774707 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-catalog-content\") pod \"certified-operators-c6nkb\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.876019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6vrb\" (UniqueName: \"kubernetes.io/projected/35ec5155-43ef-4861-a925-28e3688546fe-kube-api-access-c6vrb\") pod \"certified-operators-c6nkb\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.876092 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-utilities\") pod \"certified-operators-c6nkb\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.876180 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-catalog-content\") pod \"certified-operators-c6nkb\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.876784 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-catalog-content\") pod \"certified-operators-c6nkb\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.876833 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-utilities\") pod \"certified-operators-c6nkb\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.902206 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6vrb\" (UniqueName: \"kubernetes.io/projected/35ec5155-43ef-4861-a925-28e3688546fe-kube-api-access-c6vrb\") pod \"certified-operators-c6nkb\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:43 crc kubenswrapper[4744]: I0106 14:48:43.992736 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:44 crc kubenswrapper[4744]: I0106 14:48:44.256518 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c6nkb"] Jan 06 14:48:44 crc kubenswrapper[4744]: I0106 14:48:44.287701 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6nkb" event={"ID":"35ec5155-43ef-4861-a925-28e3688546fe","Type":"ContainerStarted","Data":"0221b8bd3b23109a1d1d5b5c1e2c60699e57d94df7c37b15e73d77115e9d5921"} Jan 06 14:48:44 crc kubenswrapper[4744]: I0106 14:48:44.423747 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:48:44 crc kubenswrapper[4744]: I0106 14:48:44.423805 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:48:45 crc kubenswrapper[4744]: I0106 14:48:45.294913 4744 generic.go:334] "Generic (PLEG): container finished" podID="35ec5155-43ef-4861-a925-28e3688546fe" containerID="94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec" exitCode=0 Jan 06 14:48:45 crc kubenswrapper[4744]: I0106 14:48:45.294995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6nkb" event={"ID":"35ec5155-43ef-4861-a925-28e3688546fe","Type":"ContainerDied","Data":"94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec"} Jan 06 14:48:46 crc kubenswrapper[4744]: I0106 14:48:46.304588 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck" event={"ID":"8a59d77a-d5d2-4a5c-a63d-cdacacf67f0b","Type":"ContainerStarted","Data":"6857c62916529c3fde76cec40263946b261ed581885e6596f2a9ca865dec2a6b"} Jan 06 14:48:46 crc kubenswrapper[4744]: I0106 14:48:46.340102 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-4qxck" podStartSLOduration=2.0465649949999998 podStartE2EDuration="32.340081951s" podCreationTimestamp="2026-01-06 14:48:14 +0000 UTC" firstStartedPulling="2026-01-06 14:48:15.788015616 +0000 UTC m=+692.415481934" lastFinishedPulling="2026-01-06 14:48:46.081532532 +0000 UTC m=+722.708998890" observedRunningTime="2026-01-06 14:48:46.337822851 +0000 UTC m=+722.965289169" watchObservedRunningTime="2026-01-06 14:48:46.340081951 +0000 UTC m=+722.967548279" Jan 06 14:48:47 crc kubenswrapper[4744]: I0106 14:48:47.313564 4744 generic.go:334] "Generic (PLEG): container finished" podID="35ec5155-43ef-4861-a925-28e3688546fe" containerID="13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3" exitCode=0 Jan 06 14:48:47 crc kubenswrapper[4744]: I0106 14:48:47.313648 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6nkb" event={"ID":"35ec5155-43ef-4861-a925-28e3688546fe","Type":"ContainerDied","Data":"13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3"} Jan 06 14:48:48 crc kubenswrapper[4744]: I0106 14:48:48.321836 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6nkb" event={"ID":"35ec5155-43ef-4861-a925-28e3688546fe","Type":"ContainerStarted","Data":"6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62"} Jan 06 14:48:48 crc kubenswrapper[4744]: I0106 14:48:48.342524 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c6nkb" podStartSLOduration=3.36883229 podStartE2EDuration="5.342508061s" podCreationTimestamp="2026-01-06 14:48:43 +0000 UTC" firstStartedPulling="2026-01-06 14:48:46.004757664 +0000 UTC m=+722.632223982" lastFinishedPulling="2026-01-06 14:48:47.978433405 +0000 UTC m=+724.605899753" observedRunningTime="2026-01-06 14:48:48.337864428 +0000 UTC m=+724.965330756" watchObservedRunningTime="2026-01-06 14:48:48.342508061 +0000 UTC m=+724.969974379" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.333549 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.334887 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.338683 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.338687 4744 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-2mfxd" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.338984 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.357878 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.495623 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-00253ab1-fc6e-4b7d-8d4c-11ae03cb48fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00253ab1-fc6e-4b7d-8d4c-11ae03cb48fc\") pod \"minio\" (UID: \"dd3ab103-081a-43f1-b40f-ae3b18862735\") " pod="minio-dev/minio" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.495697 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr4fh\" (UniqueName: \"kubernetes.io/projected/dd3ab103-081a-43f1-b40f-ae3b18862735-kube-api-access-qr4fh\") pod \"minio\" (UID: \"dd3ab103-081a-43f1-b40f-ae3b18862735\") " pod="minio-dev/minio" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.597377 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-00253ab1-fc6e-4b7d-8d4c-11ae03cb48fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00253ab1-fc6e-4b7d-8d4c-11ae03cb48fc\") pod \"minio\" (UID: \"dd3ab103-081a-43f1-b40f-ae3b18862735\") " pod="minio-dev/minio" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.597530 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr4fh\" (UniqueName: \"kubernetes.io/projected/dd3ab103-081a-43f1-b40f-ae3b18862735-kube-api-access-qr4fh\") pod \"minio\" (UID: \"dd3ab103-081a-43f1-b40f-ae3b18862735\") " pod="minio-dev/minio" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.602076 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.602135 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-00253ab1-fc6e-4b7d-8d4c-11ae03cb48fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00253ab1-fc6e-4b7d-8d4c-11ae03cb48fc\") pod \"minio\" (UID: \"dd3ab103-081a-43f1-b40f-ae3b18862735\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bf61817cf6b1b9ae1800d631f81afbc25ab3994ceb138a871e1aea1290a1c86f/globalmount\"" pod="minio-dev/minio" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.624228 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-00253ab1-fc6e-4b7d-8d4c-11ae03cb48fc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00253ab1-fc6e-4b7d-8d4c-11ae03cb48fc\") pod \"minio\" (UID: \"dd3ab103-081a-43f1-b40f-ae3b18862735\") " pod="minio-dev/minio" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.629631 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr4fh\" (UniqueName: \"kubernetes.io/projected/dd3ab103-081a-43f1-b40f-ae3b18862735-kube-api-access-qr4fh\") pod \"minio\" (UID: \"dd3ab103-081a-43f1-b40f-ae3b18862735\") " pod="minio-dev/minio" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.659447 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Jan 06 14:48:51 crc kubenswrapper[4744]: I0106 14:48:51.958746 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Jan 06 14:48:51 crc kubenswrapper[4744]: W0106 14:48:51.970305 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd3ab103_081a_43f1_b40f_ae3b18862735.slice/crio-1efe0a7b2093b3e25182adbd935176861b31525eaee58758c202ed26a656d0d4 WatchSource:0}: Error finding container 1efe0a7b2093b3e25182adbd935176861b31525eaee58758c202ed26a656d0d4: Status 404 returned error can't find the container with id 1efe0a7b2093b3e25182adbd935176861b31525eaee58758c202ed26a656d0d4 Jan 06 14:48:52 crc kubenswrapper[4744]: I0106 14:48:52.375442 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"dd3ab103-081a-43f1-b40f-ae3b18862735","Type":"ContainerStarted","Data":"1efe0a7b2093b3e25182adbd935176861b31525eaee58758c202ed26a656d0d4"} Jan 06 14:48:53 crc kubenswrapper[4744]: I0106 14:48:53.993301 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:53 crc kubenswrapper[4744]: I0106 14:48:53.993570 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:54 crc kubenswrapper[4744]: I0106 14:48:54.042385 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:54 crc kubenswrapper[4744]: I0106 14:48:54.433930 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:54 crc kubenswrapper[4744]: I0106 14:48:54.489365 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c6nkb"] Jan 06 14:48:56 crc kubenswrapper[4744]: I0106 14:48:56.406652 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"dd3ab103-081a-43f1-b40f-ae3b18862735","Type":"ContainerStarted","Data":"c816ec0a04e3210335e213821d0ef2559ac27c1d465ea1572c83f3e4ba779239"} Jan 06 14:48:56 crc kubenswrapper[4744]: I0106 14:48:56.407276 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c6nkb" podUID="35ec5155-43ef-4861-a925-28e3688546fe" containerName="registry-server" containerID="cri-o://6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62" gracePeriod=2 Jan 06 14:48:56 crc kubenswrapper[4744]: I0106 14:48:56.449624 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.312723157 podStartE2EDuration="8.449600326s" podCreationTimestamp="2026-01-06 14:48:48 +0000 UTC" firstStartedPulling="2026-01-06 14:48:51.971828793 +0000 UTC m=+728.599295111" lastFinishedPulling="2026-01-06 14:48:56.108705962 +0000 UTC m=+732.736172280" observedRunningTime="2026-01-06 14:48:56.43842188 +0000 UTC m=+733.065888238" watchObservedRunningTime="2026-01-06 14:48:56.449600326 +0000 UTC m=+733.077066674" Jan 06 14:48:56 crc kubenswrapper[4744]: I0106 14:48:56.831591 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.002079 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6vrb\" (UniqueName: \"kubernetes.io/projected/35ec5155-43ef-4861-a925-28e3688546fe-kube-api-access-c6vrb\") pod \"35ec5155-43ef-4861-a925-28e3688546fe\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.002221 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-catalog-content\") pod \"35ec5155-43ef-4861-a925-28e3688546fe\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.002273 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-utilities\") pod \"35ec5155-43ef-4861-a925-28e3688546fe\" (UID: \"35ec5155-43ef-4861-a925-28e3688546fe\") " Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.004187 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-utilities" (OuterVolumeSpecName: "utilities") pod "35ec5155-43ef-4861-a925-28e3688546fe" (UID: "35ec5155-43ef-4861-a925-28e3688546fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.016446 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35ec5155-43ef-4861-a925-28e3688546fe-kube-api-access-c6vrb" (OuterVolumeSpecName: "kube-api-access-c6vrb") pod "35ec5155-43ef-4861-a925-28e3688546fe" (UID: "35ec5155-43ef-4861-a925-28e3688546fe"). InnerVolumeSpecName "kube-api-access-c6vrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.095731 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35ec5155-43ef-4861-a925-28e3688546fe" (UID: "35ec5155-43ef-4861-a925-28e3688546fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.104819 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.104889 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6vrb\" (UniqueName: \"kubernetes.io/projected/35ec5155-43ef-4861-a925-28e3688546fe-kube-api-access-c6vrb\") on node \"crc\" DevicePath \"\"" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.104918 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35ec5155-43ef-4861-a925-28e3688546fe-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.419490 4744 generic.go:334] "Generic (PLEG): container finished" podID="35ec5155-43ef-4861-a925-28e3688546fe" containerID="6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62" exitCode=0 Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.419557 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6nkb" event={"ID":"35ec5155-43ef-4861-a925-28e3688546fe","Type":"ContainerDied","Data":"6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62"} Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.419614 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c6nkb" event={"ID":"35ec5155-43ef-4861-a925-28e3688546fe","Type":"ContainerDied","Data":"0221b8bd3b23109a1d1d5b5c1e2c60699e57d94df7c37b15e73d77115e9d5921"} Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.419633 4744 scope.go:117] "RemoveContainer" containerID="6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.419570 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c6nkb" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.454521 4744 scope.go:117] "RemoveContainer" containerID="13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.459333 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c6nkb"] Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.470345 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c6nkb"] Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.493747 4744 scope.go:117] "RemoveContainer" containerID="94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.525210 4744 scope.go:117] "RemoveContainer" containerID="6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62" Jan 06 14:48:57 crc kubenswrapper[4744]: E0106 14:48:57.525951 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62\": container with ID starting with 6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62 not found: ID does not exist" containerID="6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.526000 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62"} err="failed to get container status \"6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62\": rpc error: code = NotFound desc = could not find container \"6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62\": container with ID starting with 6f75a04ec2e35878063151030c06b4e98c33267125e4520c924b074743e7bf62 not found: ID does not exist" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.526022 4744 scope.go:117] "RemoveContainer" containerID="13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3" Jan 06 14:48:57 crc kubenswrapper[4744]: E0106 14:48:57.526710 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3\": container with ID starting with 13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3 not found: ID does not exist" containerID="13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.526734 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3"} err="failed to get container status \"13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3\": rpc error: code = NotFound desc = could not find container \"13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3\": container with ID starting with 13d364907bfea4802ca4103a71abe2412b6f1e1aa589434a7094f07c557a39a3 not found: ID does not exist" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.526775 4744 scope.go:117] "RemoveContainer" containerID="94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec" Jan 06 14:48:57 crc kubenswrapper[4744]: E0106 14:48:57.527185 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec\": container with ID starting with 94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec not found: ID does not exist" containerID="94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.527207 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec"} err="failed to get container status \"94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec\": rpc error: code = NotFound desc = could not find container \"94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec\": container with ID starting with 94485219be5eaa213525300314dd6d3206a8122d8c7e57277851d7e4c4f80dec not found: ID does not exist" Jan 06 14:48:57 crc kubenswrapper[4744]: I0106 14:48:57.721375 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35ec5155-43ef-4861-a925-28e3688546fe" path="/var/lib/kubelet/pods/35ec5155-43ef-4861-a925-28e3688546fe/volumes" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.857012 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9"] Jan 06 14:49:01 crc kubenswrapper[4744]: E0106 14:49:01.857534 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ec5155-43ef-4861-a925-28e3688546fe" containerName="extract-content" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.857548 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ec5155-43ef-4861-a925-28e3688546fe" containerName="extract-content" Jan 06 14:49:01 crc kubenswrapper[4744]: E0106 14:49:01.857575 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ec5155-43ef-4861-a925-28e3688546fe" containerName="extract-utilities" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.857586 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ec5155-43ef-4861-a925-28e3688546fe" containerName="extract-utilities" Jan 06 14:49:01 crc kubenswrapper[4744]: E0106 14:49:01.857602 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ec5155-43ef-4861-a925-28e3688546fe" containerName="registry-server" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.857611 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ec5155-43ef-4861-a925-28e3688546fe" containerName="registry-server" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.857757 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="35ec5155-43ef-4861-a925-28e3688546fe" containerName="registry-server" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.858298 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.862688 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.862713 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.862895 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-8qmd9" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.863058 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.863112 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.876581 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9"] Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.976886 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7549f443-f7d9-42fe-94a0-5bc9565caa14-logging-loki-ca-bundle\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.976938 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7549f443-f7d9-42fe-94a0-5bc9565caa14-config\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.976968 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvsks\" (UniqueName: \"kubernetes.io/projected/7549f443-f7d9-42fe-94a0-5bc9565caa14-kube-api-access-gvsks\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.977028 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/7549f443-f7d9-42fe-94a0-5bc9565caa14-logging-loki-distributor-http\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:01 crc kubenswrapper[4744]: I0106 14:49:01.977088 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/7549f443-f7d9-42fe-94a0-5bc9565caa14-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.010666 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-76788598db-v9bbz"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.011712 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.013752 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.013930 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.014186 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.038397 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-76788598db-v9bbz"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.078892 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvsks\" (UniqueName: \"kubernetes.io/projected/7549f443-f7d9-42fe-94a0-5bc9565caa14-kube-api-access-gvsks\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.079295 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/7549f443-f7d9-42fe-94a0-5bc9565caa14-logging-loki-distributor-http\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.079373 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/7549f443-f7d9-42fe-94a0-5bc9565caa14-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.079419 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7549f443-f7d9-42fe-94a0-5bc9565caa14-logging-loki-ca-bundle\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.079445 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7549f443-f7d9-42fe-94a0-5bc9565caa14-config\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.080390 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7549f443-f7d9-42fe-94a0-5bc9565caa14-config\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.081087 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7549f443-f7d9-42fe-94a0-5bc9565caa14-logging-loki-ca-bundle\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.084138 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.088406 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.095645 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.095824 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.096654 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/7549f443-f7d9-42fe-94a0-5bc9565caa14-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.102455 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.104720 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/7549f443-f7d9-42fe-94a0-5bc9565caa14-logging-loki-distributor-http\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.122922 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvsks\" (UniqueName: \"kubernetes.io/projected/7549f443-f7d9-42fe-94a0-5bc9565caa14-kube-api-access-gvsks\") pod \"logging-loki-distributor-5f678c8dd6-jtrh9\" (UID: \"7549f443-f7d9-42fe-94a0-5bc9565caa14\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.176450 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.179550 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-b67674996-jsskr"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180051 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180210 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2c75046-2bab-4e58-b54a-0172540755a2-config\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180234 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtstt\" (UniqueName: \"kubernetes.io/projected/a2c75046-2bab-4e58-b54a-0172540755a2-kube-api-access-wtstt\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180268 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnpf4\" (UniqueName: \"kubernetes.io/projected/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-kube-api-access-cnpf4\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180333 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-config\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180366 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-querier-http\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-querier-grpc\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180485 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180518 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180494 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180558 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-s3\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.180680 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-ca-bundle\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.184399 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.184517 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.184545 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.184544 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.184447 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.192545 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-b67674996-w4jmm"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.204881 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.217180 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-rjp4q" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.246558 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-b67674996-jsskr"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.258064 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-b67674996-w4jmm"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281701 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-s3\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281765 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281795 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-ca-bundle\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281817 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-tls-secret\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281839 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281857 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-lokistack-gateway\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281880 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2c75046-2bab-4e58-b54a-0172540755a2-config\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281897 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtstt\" (UniqueName: \"kubernetes.io/projected/a2c75046-2bab-4e58-b54a-0172540755a2-kube-api-access-wtstt\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281914 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-rbac\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281932 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnpf4\" (UniqueName: \"kubernetes.io/projected/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-kube-api-access-cnpf4\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281956 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-config\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-querier-http\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.281995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-querier-grpc\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.282010 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq8n7\" (UniqueName: \"kubernetes.io/projected/ddffe4cb-6ab7-4556-aa91-620d374028db-kube-api-access-wq8n7\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.282033 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.282062 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-tenants\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.282082 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.282098 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.282115 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.283442 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-ca-bundle\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.286307 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.286914 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2c75046-2bab-4e58-b54a-0172540755a2-config\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.287704 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-config\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.288725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-s3\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.290229 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-querier-http\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.292693 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.295478 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/a2c75046-2bab-4e58-b54a-0172540755a2-logging-loki-querier-grpc\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.305399 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtstt\" (UniqueName: \"kubernetes.io/projected/a2c75046-2bab-4e58-b54a-0172540755a2-kube-api-access-wtstt\") pod \"logging-loki-querier-76788598db-v9bbz\" (UID: \"a2c75046-2bab-4e58-b54a-0172540755a2\") " pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.308089 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.313371 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnpf4\" (UniqueName: \"kubernetes.io/projected/e2a5e0ac-871e-456a-9fcb-04b991a7f6d6-kube-api-access-cnpf4\") pod \"logging-loki-query-frontend-69d9546745-xbtdk\" (UID: \"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.342980 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-rbac\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383288 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-tls-secret\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wq8n7\" (UniqueName: \"kubernetes.io/projected/ddffe4cb-6ab7-4556-aa91-620d374028db-kube-api-access-wq8n7\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383346 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-tenants\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383374 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383406 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-tenants\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383466 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-lokistack-gateway\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383485 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383514 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftrb2\" (UniqueName: \"kubernetes.io/projected/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-kube-api-access-ftrb2\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383535 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-rbac\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383560 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383585 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-tls-secret\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383603 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.383624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-lokistack-gateway\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.384470 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-lokistack-gateway\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.385059 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-rbac\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.385821 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: E0106 14:49:02.386410 4744 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Jan 06 14:49:02 crc kubenswrapper[4744]: E0106 14:49:02.386555 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-tls-secret podName:ddffe4cb-6ab7-4556-aa91-620d374028db nodeName:}" failed. No retries permitted until 2026-01-06 14:49:02.886534249 +0000 UTC m=+739.514000567 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-tls-secret") pod "logging-loki-gateway-b67674996-jsskr" (UID: "ddffe4cb-6ab7-4556-aa91-620d374028db") : secret "logging-loki-gateway-http" not found Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.386905 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ddffe4cb-6ab7-4556-aa91-620d374028db-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.388722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-tenants\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.389310 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.402176 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq8n7\" (UniqueName: \"kubernetes.io/projected/ddffe4cb-6ab7-4556-aa91-620d374028db-kube-api-access-wq8n7\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.475568 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.484436 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.484493 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-tls-secret\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.484536 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-tenants\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.484565 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.484584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.484601 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-lokistack-gateway\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.484633 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftrb2\" (UniqueName: \"kubernetes.io/projected/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-kube-api-access-ftrb2\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.484652 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-rbac\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.485266 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.485373 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-rbac\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.485694 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-logging-loki-ca-bundle\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: E0106 14:49:02.485704 4744 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Jan 06 14:49:02 crc kubenswrapper[4744]: E0106 14:49:02.485788 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-tls-secret podName:4e939cdc-a027-4c89-b0e8-f5c0dff2af63 nodeName:}" failed. No retries permitted until 2026-01-06 14:49:02.98577305 +0000 UTC m=+739.613239368 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-tls-secret") pod "logging-loki-gateway-b67674996-w4jmm" (UID: "4e939cdc-a027-4c89-b0e8-f5c0dff2af63") : secret "logging-loki-gateway-http" not found Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.485857 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-lokistack-gateway\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.490659 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-tenants\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.498105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.499698 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftrb2\" (UniqueName: \"kubernetes.io/projected/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-kube-api-access-ftrb2\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.661130 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.779388 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-76788598db-v9bbz"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.895918 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-tls-secret\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.900937 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk"] Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.901919 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/ddffe4cb-6ab7-4556-aa91-620d374028db-tls-secret\") pod \"logging-loki-gateway-b67674996-jsskr\" (UID: \"ddffe4cb-6ab7-4556-aa91-620d374028db\") " pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.997639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-tls-secret\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:02 crc kubenswrapper[4744]: I0106 14:49:02.998669 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.000633 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.001156 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4e939cdc-a027-4c89-b0e8-f5c0dff2af63-tls-secret\") pod \"logging-loki-gateway-b67674996-w4jmm\" (UID: \"4e939cdc-a027-4c89-b0e8-f5c0dff2af63\") " pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.004921 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.006449 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.008467 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.080292 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.081440 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.083702 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.083961 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.096801 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.099405 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/631bc775-d5be-445e-abd6-d6ec8afd9cf2-config\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.099468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhd5q\" (UniqueName: \"kubernetes.io/projected/631bc775-d5be-445e-abd6-d6ec8afd9cf2-kube-api-access-rhd5q\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.099501 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1d80240c-56a7-4d6d-94ea-2dc31b1c371a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1d80240c-56a7-4d6d-94ea-2dc31b1c371a\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.099539 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.099565 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fd6d44dc-cc94-471f-8ec2-86c9010e76d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd6d44dc-cc94-471f-8ec2-86c9010e76d8\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.099585 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.099712 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.099756 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.104930 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.161556 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.162278 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.165878 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.170766 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.170993 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.188272 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209104 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209168 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209197 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/631bc775-d5be-445e-abd6-d6ec8afd9cf2-config\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209268 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm5mg\" (UniqueName: \"kubernetes.io/projected/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-kube-api-access-dm5mg\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhd5q\" (UniqueName: \"kubernetes.io/projected/631bc775-d5be-445e-abd6-d6ec8afd9cf2-kube-api-access-rhd5q\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209310 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1d80240c-56a7-4d6d-94ea-2dc31b1c371a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1d80240c-56a7-4d6d-94ea-2dc31b1c371a\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209352 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fd6d44dc-cc94-471f-8ec2-86c9010e76d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd6d44dc-cc94-471f-8ec2-86c9010e76d8\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209374 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209412 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209440 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ff0726c6-a369-4ae7-abc5-96c1796a3288\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ff0726c6-a369-4ae7-abc5-96c1796a3288\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.209463 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-config\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.214493 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.221962 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.222269 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/631bc775-d5be-445e-abd6-d6ec8afd9cf2-config\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.223434 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.225397 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/631bc775-d5be-445e-abd6-d6ec8afd9cf2-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.227706 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.227762 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fd6d44dc-cc94-471f-8ec2-86c9010e76d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd6d44dc-cc94-471f-8ec2-86c9010e76d8\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/61e21c67baba4305b2ad71d3c1f817a13e97fc715f9543af92f4eb015db7376f/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.228660 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.228690 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1d80240c-56a7-4d6d-94ea-2dc31b1c371a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1d80240c-56a7-4d6d-94ea-2dc31b1c371a\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ac23a5dc31765001c5f1fb4e37894c2110bdcc17ec88b45d0b8bc89c3008f028/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.234804 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhd5q\" (UniqueName: \"kubernetes.io/projected/631bc775-d5be-445e-abd6-d6ec8afd9cf2-kube-api-access-rhd5q\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.288041 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1d80240c-56a7-4d6d-94ea-2dc31b1c371a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1d80240c-56a7-4d6d-94ea-2dc31b1c371a\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.309766 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fd6d44dc-cc94-471f-8ec2-86c9010e76d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fd6d44dc-cc94-471f-8ec2-86c9010e76d8\") pod \"logging-loki-ingester-0\" (UID: \"631bc775-d5be-445e-abd6-d6ec8afd9cf2\") " pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.310732 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.310829 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.311377 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.311427 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7d402e42-bd44-4d98-b906-9a0935b58c5f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d402e42-bd44-4d98-b906-9a0935b58c5f\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.311468 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.311564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ff0726c6-a369-4ae7-abc5-96c1796a3288\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ff0726c6-a369-4ae7-abc5-96c1796a3288\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.311596 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-config\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.311665 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.311680 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.311698 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.312672 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-config\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.312747 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm5mg\" (UniqueName: \"kubernetes.io/projected/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-kube-api-access-dm5mg\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.312778 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm5sr\" (UniqueName: \"kubernetes.io/projected/d01a1f2a-5b58-4761-956c-199f1b7560fe-kube-api-access-jm5sr\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.312809 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d01a1f2a-5b58-4761-956c-199f1b7560fe-config\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.312846 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.312902 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.313714 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.313740 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ff0726c6-a369-4ae7-abc5-96c1796a3288\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ff0726c6-a369-4ae7-abc5-96c1796a3288\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/dbcc8c3abdcb7035c5378d4ba98bddaf5381f362595369e93abf9232c07371a8/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.316119 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.316827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.316891 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.337875 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.362101 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm5mg\" (UniqueName: \"kubernetes.io/projected/a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9-kube-api-access-dm5mg\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.373418 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ff0726c6-a369-4ae7-abc5-96c1796a3288\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ff0726c6-a369-4ae7-abc5-96c1796a3288\") pod \"logging-loki-compactor-0\" (UID: \"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9\") " pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.414009 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.414056 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.414085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm5sr\" (UniqueName: \"kubernetes.io/projected/d01a1f2a-5b58-4761-956c-199f1b7560fe-kube-api-access-jm5sr\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.414106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d01a1f2a-5b58-4761-956c-199f1b7560fe-config\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.414142 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.414175 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.414209 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7d402e42-bd44-4d98-b906-9a0935b58c5f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d402e42-bd44-4d98-b906-9a0935b58c5f\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.416375 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d01a1f2a-5b58-4761-956c-199f1b7560fe-config\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.417188 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.424667 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.427184 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.427541 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/d01a1f2a-5b58-4761-956c-199f1b7560fe-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.431455 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.431502 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7d402e42-bd44-4d98-b906-9a0935b58c5f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d402e42-bd44-4d98-b906-9a0935b58c5f\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b726152a6e9f5a5b6a20b490e7f1851157e977f17096ea83d31ac4674483623f/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.454787 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm5sr\" (UniqueName: \"kubernetes.io/projected/d01a1f2a-5b58-4761-956c-199f1b7560fe-kube-api-access-jm5sr\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.479515 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7d402e42-bd44-4d98-b906-9a0935b58c5f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d402e42-bd44-4d98-b906-9a0935b58c5f\") pod \"logging-loki-index-gateway-0\" (UID: \"d01a1f2a-5b58-4761-956c-199f1b7560fe\") " pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.480012 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" event={"ID":"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6","Type":"ContainerStarted","Data":"589cf8192e617b9ac6edaf50068b42c386e88422cf24afaa00be670e0f471247"} Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.481404 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" event={"ID":"a2c75046-2bab-4e58-b54a-0172540755a2","Type":"ContainerStarted","Data":"dae71808e5dd2772564b015e7284365d288b67cd334bc03d7f7004df69fd84e3"} Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.482500 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" event={"ID":"7549f443-f7d9-42fe-94a0-5bc9565caa14","Type":"ContainerStarted","Data":"3bf57a71a678ea399f8d682dd80c92eca4f45f0061c6aaeb26c419455bd76b79"} Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.487617 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.572594 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.582331 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Jan 06 14:49:03 crc kubenswrapper[4744]: W0106 14:49:03.598514 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod631bc775_d5be_445e_abd6_d6ec8afd9cf2.slice/crio-e83f02437e096e8209ef526bb430469b4eb657a1683e65310d897b2c0d1d3d47 WatchSource:0}: Error finding container e83f02437e096e8209ef526bb430469b4eb657a1683e65310d897b2c0d1d3d47: Status 404 returned error can't find the container with id e83f02437e096e8209ef526bb430469b4eb657a1683e65310d897b2c0d1d3d47 Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.700205 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-b67674996-jsskr"] Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.725886 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-b67674996-w4jmm"] Jan 06 14:49:03 crc kubenswrapper[4744]: W0106 14:49:03.729858 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e939cdc_a027_4c89_b0e8_f5c0dff2af63.slice/crio-3957cc42ee8ad222dc79a1b51c993a2eb74695355e23a9d9ccdb8d8b281a12a5 WatchSource:0}: Error finding container 3957cc42ee8ad222dc79a1b51c993a2eb74695355e23a9d9ccdb8d8b281a12a5: Status 404 returned error can't find the container with id 3957cc42ee8ad222dc79a1b51c993a2eb74695355e23a9d9ccdb8d8b281a12a5 Jan 06 14:49:03 crc kubenswrapper[4744]: I0106 14:49:03.909441 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Jan 06 14:49:03 crc kubenswrapper[4744]: W0106 14:49:03.919983 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda48ae2c1_a7b3_43e2_b7de_f2c6cafea7b9.slice/crio-1e160d7e6e282cc8c7a2a7ef9386d569f15a0e563d24689bb99f921ebde4c5fe WatchSource:0}: Error finding container 1e160d7e6e282cc8c7a2a7ef9386d569f15a0e563d24689bb99f921ebde4c5fe: Status 404 returned error can't find the container with id 1e160d7e6e282cc8c7a2a7ef9386d569f15a0e563d24689bb99f921ebde4c5fe Jan 06 14:49:04 crc kubenswrapper[4744]: I0106 14:49:04.017510 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Jan 06 14:49:04 crc kubenswrapper[4744]: W0106 14:49:04.026139 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd01a1f2a_5b58_4761_956c_199f1b7560fe.slice/crio-5d212af825c39332c7182f8691f2566fcf29976dae0326e9c361fde242a3c4a0 WatchSource:0}: Error finding container 5d212af825c39332c7182f8691f2566fcf29976dae0326e9c361fde242a3c4a0: Status 404 returned error can't find the container with id 5d212af825c39332c7182f8691f2566fcf29976dae0326e9c361fde242a3c4a0 Jan 06 14:49:04 crc kubenswrapper[4744]: I0106 14:49:04.491988 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"d01a1f2a-5b58-4761-956c-199f1b7560fe","Type":"ContainerStarted","Data":"5d212af825c39332c7182f8691f2566fcf29976dae0326e9c361fde242a3c4a0"} Jan 06 14:49:04 crc kubenswrapper[4744]: I0106 14:49:04.493372 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" event={"ID":"4e939cdc-a027-4c89-b0e8-f5c0dff2af63","Type":"ContainerStarted","Data":"3957cc42ee8ad222dc79a1b51c993a2eb74695355e23a9d9ccdb8d8b281a12a5"} Jan 06 14:49:04 crc kubenswrapper[4744]: I0106 14:49:04.494520 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" event={"ID":"ddffe4cb-6ab7-4556-aa91-620d374028db","Type":"ContainerStarted","Data":"bb6adf8ffbfacfae7dc22aec11c9b68be484115bfaac0959159dca84f693fcda"} Jan 06 14:49:04 crc kubenswrapper[4744]: I0106 14:49:04.495421 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"631bc775-d5be-445e-abd6-d6ec8afd9cf2","Type":"ContainerStarted","Data":"e83f02437e096e8209ef526bb430469b4eb657a1683e65310d897b2c0d1d3d47"} Jan 06 14:49:04 crc kubenswrapper[4744]: I0106 14:49:04.496835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9","Type":"ContainerStarted","Data":"1e160d7e6e282cc8c7a2a7ef9386d569f15a0e563d24689bb99f921ebde4c5fe"} Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.521564 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9","Type":"ContainerStarted","Data":"73770af608a58d12440b3d866c14c738c487ed3848cf894614a843c77dec272e"} Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.522093 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.526743 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"d01a1f2a-5b58-4761-956c-199f1b7560fe","Type":"ContainerStarted","Data":"88d56755db3132658ea7004374932dae3eeea73fdeda762ba7acddb1d6a7ff87"} Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.528515 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" event={"ID":"7549f443-f7d9-42fe-94a0-5bc9565caa14","Type":"ContainerStarted","Data":"39a5aee61c56e4cc41844ec2b3bf7567a184b5db5de0707e25dac633e6da7712"} Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.529207 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.529377 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.530722 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" event={"ID":"e2a5e0ac-871e-456a-9fcb-04b991a7f6d6","Type":"ContainerStarted","Data":"a9cd2823ad8e4a2055b72ea992ba6b765dd7d31bfc3d52cc85f0f3487aa84a87"} Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.531230 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.533057 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" event={"ID":"a2c75046-2bab-4e58-b54a-0172540755a2","Type":"ContainerStarted","Data":"14e6fd7166d1e96413e7d5777a2c2bcabf0aa82d9ad98e2bf36f5c1512d7845d"} Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.533661 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.535350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"631bc775-d5be-445e-abd6-d6ec8afd9cf2","Type":"ContainerStarted","Data":"5f29fe673a4682b7db0e38aaf1be34462e86c81a4510d825dcc0bc05a45e6283"} Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.535918 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.574373 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" podStartSLOduration=2.817626438 podStartE2EDuration="5.574349863s" podCreationTimestamp="2026-01-06 14:49:01 +0000 UTC" firstStartedPulling="2026-01-06 14:49:02.692357947 +0000 UTC m=+739.319824265" lastFinishedPulling="2026-01-06 14:49:05.449081372 +0000 UTC m=+742.076547690" observedRunningTime="2026-01-06 14:49:06.570904322 +0000 UTC m=+743.198370640" watchObservedRunningTime="2026-01-06 14:49:06.574349863 +0000 UTC m=+743.201816181" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.574825 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=2.905122764 podStartE2EDuration="4.574817105s" podCreationTimestamp="2026-01-06 14:49:02 +0000 UTC" firstStartedPulling="2026-01-06 14:49:03.923740442 +0000 UTC m=+740.551206770" lastFinishedPulling="2026-01-06 14:49:05.593434793 +0000 UTC m=+742.220901111" observedRunningTime="2026-01-06 14:49:06.548630934 +0000 UTC m=+743.176097282" watchObservedRunningTime="2026-01-06 14:49:06.574817105 +0000 UTC m=+743.202283423" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.592680 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=3.710597945 podStartE2EDuration="5.592665487s" podCreationTimestamp="2026-01-06 14:49:01 +0000 UTC" firstStartedPulling="2026-01-06 14:49:03.602191629 +0000 UTC m=+740.229657937" lastFinishedPulling="2026-01-06 14:49:05.484259171 +0000 UTC m=+742.111725479" observedRunningTime="2026-01-06 14:49:06.591019373 +0000 UTC m=+743.218485691" watchObservedRunningTime="2026-01-06 14:49:06.592665487 +0000 UTC m=+743.220131805" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.612777 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" podStartSLOduration=2.945221558 podStartE2EDuration="5.612759117s" podCreationTimestamp="2026-01-06 14:49:01 +0000 UTC" firstStartedPulling="2026-01-06 14:49:02.781144722 +0000 UTC m=+739.408611040" lastFinishedPulling="2026-01-06 14:49:05.448682281 +0000 UTC m=+742.076148599" observedRunningTime="2026-01-06 14:49:06.607123399 +0000 UTC m=+743.234589727" watchObservedRunningTime="2026-01-06 14:49:06.612759117 +0000 UTC m=+743.240225435" Jan 06 14:49:06 crc kubenswrapper[4744]: I0106 14:49:06.634996 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.121719065 podStartE2EDuration="4.634957714s" podCreationTimestamp="2026-01-06 14:49:02 +0000 UTC" firstStartedPulling="2026-01-06 14:49:04.030026629 +0000 UTC m=+740.657492947" lastFinishedPulling="2026-01-06 14:49:05.543265278 +0000 UTC m=+742.170731596" observedRunningTime="2026-01-06 14:49:06.627390004 +0000 UTC m=+743.254856322" watchObservedRunningTime="2026-01-06 14:49:06.634957714 +0000 UTC m=+743.262424032" Jan 06 14:49:07 crc kubenswrapper[4744]: I0106 14:49:07.550643 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" event={"ID":"4e939cdc-a027-4c89-b0e8-f5c0dff2af63","Type":"ContainerStarted","Data":"cda853a9373f27f441c2dbfe56af764f2deb2b8629282c8523dfb7759bb03145"} Jan 06 14:49:07 crc kubenswrapper[4744]: I0106 14:49:07.553367 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" event={"ID":"ddffe4cb-6ab7-4556-aa91-620d374028db","Type":"ContainerStarted","Data":"901d28b6520a3de955d1ffb1acc96cad25dc708a6a4748fcfdc6911ae185bd18"} Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.590224 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" event={"ID":"4e939cdc-a027-4c89-b0e8-f5c0dff2af63","Type":"ContainerStarted","Data":"22af061a2195290e675c6a425887d476ada53d76c66130db42e7010622cbfe3d"} Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.590780 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.590816 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.594546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" event={"ID":"ddffe4cb-6ab7-4556-aa91-620d374028db","Type":"ContainerStarted","Data":"897579a31a294122f1caadb551e07346864b43ae5d7e340f6cb5347e9f292ed9"} Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.595743 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.595844 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.614626 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.615954 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.618830 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.619572 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.628101 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" podStartSLOduration=2.647951361 podStartE2EDuration="8.628087575s" podCreationTimestamp="2026-01-06 14:49:02 +0000 UTC" firstStartedPulling="2026-01-06 14:49:03.736483176 +0000 UTC m=+740.363949494" lastFinishedPulling="2026-01-06 14:49:09.71661938 +0000 UTC m=+746.344085708" observedRunningTime="2026-01-06 14:49:10.62597778 +0000 UTC m=+747.253444158" watchObservedRunningTime="2026-01-06 14:49:10.628087575 +0000 UTC m=+747.255553903" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.631777 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" podStartSLOduration=5.997222817 podStartE2EDuration="8.631768753s" podCreationTimestamp="2026-01-06 14:49:02 +0000 UTC" firstStartedPulling="2026-01-06 14:49:02.90599257 +0000 UTC m=+739.533458888" lastFinishedPulling="2026-01-06 14:49:05.540538506 +0000 UTC m=+742.168004824" observedRunningTime="2026-01-06 14:49:06.651340676 +0000 UTC m=+743.278807024" watchObservedRunningTime="2026-01-06 14:49:10.631768753 +0000 UTC m=+747.259235081" Jan 06 14:49:10 crc kubenswrapper[4744]: I0106 14:49:10.666274 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-b67674996-jsskr" podStartSLOduration=2.66003206 podStartE2EDuration="8.666250643s" podCreationTimestamp="2026-01-06 14:49:02 +0000 UTC" firstStartedPulling="2026-01-06 14:49:03.702215491 +0000 UTC m=+740.329681809" lastFinishedPulling="2026-01-06 14:49:09.708434054 +0000 UTC m=+746.335900392" observedRunningTime="2026-01-06 14:49:10.661631771 +0000 UTC m=+747.289098119" watchObservedRunningTime="2026-01-06 14:49:10.666250643 +0000 UTC m=+747.293716981" Jan 06 14:49:14 crc kubenswrapper[4744]: I0106 14:49:14.423877 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:49:14 crc kubenswrapper[4744]: I0106 14:49:14.424911 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:49:14 crc kubenswrapper[4744]: I0106 14:49:14.424983 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:49:14 crc kubenswrapper[4744]: I0106 14:49:14.426051 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"098e8945831f6c03a5214e0dd27553b61625d63e6339c0ac5bf8762f7f51444d"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 14:49:14 crc kubenswrapper[4744]: I0106 14:49:14.426130 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://098e8945831f6c03a5214e0dd27553b61625d63e6339c0ac5bf8762f7f51444d" gracePeriod=600 Jan 06 14:49:14 crc kubenswrapper[4744]: I0106 14:49:14.622394 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="098e8945831f6c03a5214e0dd27553b61625d63e6339c0ac5bf8762f7f51444d" exitCode=0 Jan 06 14:49:14 crc kubenswrapper[4744]: I0106 14:49:14.622444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"098e8945831f6c03a5214e0dd27553b61625d63e6339c0ac5bf8762f7f51444d"} Jan 06 14:49:14 crc kubenswrapper[4744]: I0106 14:49:14.622482 4744 scope.go:117] "RemoveContainer" containerID="c86abf51f54161b2a9bee59a750138794b41f57e70d71c761e86483f896721fe" Jan 06 14:49:15 crc kubenswrapper[4744]: I0106 14:49:15.637345 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"b7fdf0a873dd4feb170b2380a6eb8f3f910485e646fb343c689eab9c501fa171"} Jan 06 14:49:22 crc kubenswrapper[4744]: I0106 14:49:22.188192 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-jtrh9" Jan 06 14:49:22 crc kubenswrapper[4744]: I0106 14:49:22.358382 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-76788598db-v9bbz" Jan 06 14:49:22 crc kubenswrapper[4744]: I0106 14:49:22.574548 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-69d9546745-xbtdk" Jan 06 14:49:23 crc kubenswrapper[4744]: I0106 14:49:23.348489 4744 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Jan 06 14:49:23 crc kubenswrapper[4744]: I0106 14:49:23.348590 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="631bc775-d5be-445e-abd6-d6ec8afd9cf2" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 06 14:49:23 crc kubenswrapper[4744]: I0106 14:49:23.500646 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Jan 06 14:49:23 crc kubenswrapper[4744]: I0106 14:49:23.595974 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Jan 06 14:49:33 crc kubenswrapper[4744]: I0106 14:49:33.344943 4744 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Jan 06 14:49:33 crc kubenswrapper[4744]: I0106 14:49:33.345735 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="631bc775-d5be-445e-abd6-d6ec8afd9cf2" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 06 14:49:43 crc kubenswrapper[4744]: I0106 14:49:43.346507 4744 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Jan 06 14:49:43 crc kubenswrapper[4744]: I0106 14:49:43.347391 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="631bc775-d5be-445e-abd6-d6ec8afd9cf2" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.011350 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-crgnf"] Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.013681 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.022799 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-crgnf"] Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.139033 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-catalog-content\") pod \"community-operators-crgnf\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.139457 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zqqx\" (UniqueName: \"kubernetes.io/projected/3e413d35-2691-4391-a6b5-73b3acbeaefd-kube-api-access-7zqqx\") pod \"community-operators-crgnf\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.139578 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-utilities\") pod \"community-operators-crgnf\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.241896 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-utilities\") pod \"community-operators-crgnf\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.242031 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-catalog-content\") pod \"community-operators-crgnf\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.242249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zqqx\" (UniqueName: \"kubernetes.io/projected/3e413d35-2691-4391-a6b5-73b3acbeaefd-kube-api-access-7zqqx\") pod \"community-operators-crgnf\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.243212 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-catalog-content\") pod \"community-operators-crgnf\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.243213 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-utilities\") pod \"community-operators-crgnf\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.265294 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zqqx\" (UniqueName: \"kubernetes.io/projected/3e413d35-2691-4391-a6b5-73b3acbeaefd-kube-api-access-7zqqx\") pod \"community-operators-crgnf\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.354009 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.617031 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-crgnf"] Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.928447 4744 generic.go:334] "Generic (PLEG): container finished" podID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerID="a4df629f25b425fdcb528b55c1758dbb6b149075475626ad70bbbdf6f31a1235" exitCode=0 Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.928510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crgnf" event={"ID":"3e413d35-2691-4391-a6b5-73b3acbeaefd","Type":"ContainerDied","Data":"a4df629f25b425fdcb528b55c1758dbb6b149075475626ad70bbbdf6f31a1235"} Jan 06 14:49:45 crc kubenswrapper[4744]: I0106 14:49:45.928551 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crgnf" event={"ID":"3e413d35-2691-4391-a6b5-73b3acbeaefd","Type":"ContainerStarted","Data":"83ac7bd9fa90c176181cb95b7b5ce8d1d480659760573703edc6e5e529a374ba"} Jan 06 14:49:46 crc kubenswrapper[4744]: I0106 14:49:46.940040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crgnf" event={"ID":"3e413d35-2691-4391-a6b5-73b3acbeaefd","Type":"ContainerStarted","Data":"6f9cd16fb23ac041d3a94c88a3dc534e76b6f047a4d16b1e1ba4d3ddbdf584e0"} Jan 06 14:49:47 crc kubenswrapper[4744]: I0106 14:49:47.964918 4744 generic.go:334] "Generic (PLEG): container finished" podID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerID="6f9cd16fb23ac041d3a94c88a3dc534e76b6f047a4d16b1e1ba4d3ddbdf584e0" exitCode=0 Jan 06 14:49:47 crc kubenswrapper[4744]: I0106 14:49:47.964973 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crgnf" event={"ID":"3e413d35-2691-4391-a6b5-73b3acbeaefd","Type":"ContainerDied","Data":"6f9cd16fb23ac041d3a94c88a3dc534e76b6f047a4d16b1e1ba4d3ddbdf584e0"} Jan 06 14:49:48 crc kubenswrapper[4744]: I0106 14:49:48.974818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crgnf" event={"ID":"3e413d35-2691-4391-a6b5-73b3acbeaefd","Type":"ContainerStarted","Data":"1fdae1263c63b51153bce6be24d3309066dd65c8d74a05ddac2d0281d698c5da"} Jan 06 14:49:49 crc kubenswrapper[4744]: I0106 14:49:48.999957 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-crgnf" podStartSLOduration=2.530538006 podStartE2EDuration="4.999932219s" podCreationTimestamp="2026-01-06 14:49:44 +0000 UTC" firstStartedPulling="2026-01-06 14:49:45.932567847 +0000 UTC m=+782.560034195" lastFinishedPulling="2026-01-06 14:49:48.40196208 +0000 UTC m=+785.029428408" observedRunningTime="2026-01-06 14:49:48.99844559 +0000 UTC m=+785.625911938" watchObservedRunningTime="2026-01-06 14:49:48.999932219 +0000 UTC m=+785.627398567" Jan 06 14:49:53 crc kubenswrapper[4744]: I0106 14:49:53.344886 4744 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Jan 06 14:49:53 crc kubenswrapper[4744]: I0106 14:49:53.345479 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="631bc775-d5be-445e-abd6-d6ec8afd9cf2" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 06 14:49:55 crc kubenswrapper[4744]: I0106 14:49:55.354542 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:55 crc kubenswrapper[4744]: I0106 14:49:55.355025 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:55 crc kubenswrapper[4744]: I0106 14:49:55.423423 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:56 crc kubenswrapper[4744]: I0106 14:49:56.112962 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:56 crc kubenswrapper[4744]: I0106 14:49:56.180579 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-crgnf"] Jan 06 14:49:58 crc kubenswrapper[4744]: I0106 14:49:58.059223 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-crgnf" podUID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerName="registry-server" containerID="cri-o://1fdae1263c63b51153bce6be24d3309066dd65c8d74a05ddac2d0281d698c5da" gracePeriod=2 Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.068812 4744 generic.go:334] "Generic (PLEG): container finished" podID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerID="1fdae1263c63b51153bce6be24d3309066dd65c8d74a05ddac2d0281d698c5da" exitCode=0 Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.068907 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crgnf" event={"ID":"3e413d35-2691-4391-a6b5-73b3acbeaefd","Type":"ContainerDied","Data":"1fdae1263c63b51153bce6be24d3309066dd65c8d74a05ddac2d0281d698c5da"} Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.168995 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.307889 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zqqx\" (UniqueName: \"kubernetes.io/projected/3e413d35-2691-4391-a6b5-73b3acbeaefd-kube-api-access-7zqqx\") pod \"3e413d35-2691-4391-a6b5-73b3acbeaefd\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.307960 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-catalog-content\") pod \"3e413d35-2691-4391-a6b5-73b3acbeaefd\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.308010 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-utilities\") pod \"3e413d35-2691-4391-a6b5-73b3acbeaefd\" (UID: \"3e413d35-2691-4391-a6b5-73b3acbeaefd\") " Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.309557 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-utilities" (OuterVolumeSpecName: "utilities") pod "3e413d35-2691-4391-a6b5-73b3acbeaefd" (UID: "3e413d35-2691-4391-a6b5-73b3acbeaefd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.313949 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e413d35-2691-4391-a6b5-73b3acbeaefd-kube-api-access-7zqqx" (OuterVolumeSpecName: "kube-api-access-7zqqx") pod "3e413d35-2691-4391-a6b5-73b3acbeaefd" (UID: "3e413d35-2691-4391-a6b5-73b3acbeaefd"). InnerVolumeSpecName "kube-api-access-7zqqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.392189 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e413d35-2691-4391-a6b5-73b3acbeaefd" (UID: "3e413d35-2691-4391-a6b5-73b3acbeaefd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.409467 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zqqx\" (UniqueName: \"kubernetes.io/projected/3e413d35-2691-4391-a6b5-73b3acbeaefd-kube-api-access-7zqqx\") on node \"crc\" DevicePath \"\"" Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.409511 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:49:59 crc kubenswrapper[4744]: I0106 14:49:59.409524 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e413d35-2691-4391-a6b5-73b3acbeaefd-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:00 crc kubenswrapper[4744]: I0106 14:50:00.080259 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crgnf" event={"ID":"3e413d35-2691-4391-a6b5-73b3acbeaefd","Type":"ContainerDied","Data":"83ac7bd9fa90c176181cb95b7b5ce8d1d480659760573703edc6e5e529a374ba"} Jan 06 14:50:00 crc kubenswrapper[4744]: I0106 14:50:00.080307 4744 scope.go:117] "RemoveContainer" containerID="1fdae1263c63b51153bce6be24d3309066dd65c8d74a05ddac2d0281d698c5da" Jan 06 14:50:00 crc kubenswrapper[4744]: I0106 14:50:00.080361 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crgnf" Jan 06 14:50:00 crc kubenswrapper[4744]: I0106 14:50:00.102843 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-crgnf"] Jan 06 14:50:00 crc kubenswrapper[4744]: I0106 14:50:00.113488 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-crgnf"] Jan 06 14:50:00 crc kubenswrapper[4744]: I0106 14:50:00.115139 4744 scope.go:117] "RemoveContainer" containerID="6f9cd16fb23ac041d3a94c88a3dc534e76b6f047a4d16b1e1ba4d3ddbdf584e0" Jan 06 14:50:00 crc kubenswrapper[4744]: I0106 14:50:00.148523 4744 scope.go:117] "RemoveContainer" containerID="a4df629f25b425fdcb528b55c1758dbb6b149075475626ad70bbbdf6f31a1235" Jan 06 14:50:01 crc kubenswrapper[4744]: I0106 14:50:01.727181 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e413d35-2691-4391-a6b5-73b3acbeaefd" path="/var/lib/kubelet/pods/3e413d35-2691-4391-a6b5-73b3acbeaefd/volumes" Jan 06 14:50:03 crc kubenswrapper[4744]: I0106 14:50:03.345896 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.759629 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-v2cmh"] Jan 06 14:50:21 crc kubenswrapper[4744]: E0106 14:50:21.760482 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerName="registry-server" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.760496 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerName="registry-server" Jan 06 14:50:21 crc kubenswrapper[4744]: E0106 14:50:21.760521 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerName="extract-utilities" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.760528 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerName="extract-utilities" Jan 06 14:50:21 crc kubenswrapper[4744]: E0106 14:50:21.760537 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerName="extract-content" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.760543 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerName="extract-content" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.760664 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e413d35-2691-4391-a6b5-73b3acbeaefd" containerName="registry-server" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.761137 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.762806 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.763976 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.764053 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-7lhp8" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.765983 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.768207 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.771464 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.789752 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-v2cmh"] Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.837637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-token\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.837976 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-datadir\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.838150 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-entrypoint\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.838327 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.838449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-sa-token\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.838551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-metrics\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.838675 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-trusted-ca\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.838801 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-syslog-receiver\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.838926 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config-openshift-service-cacrt\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.839048 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g24gl\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-kube-api-access-g24gl\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.839225 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-tmp\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.937458 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-v2cmh"] Jan 06 14:50:21 crc kubenswrapper[4744]: E0106 14:50:21.938143 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-g24gl metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-v2cmh" podUID="8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940347 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config-openshift-service-cacrt\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940412 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g24gl\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-kube-api-access-g24gl\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940441 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-tmp\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940484 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-token\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-datadir\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-entrypoint\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940574 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-sa-token\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-metrics\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940655 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-datadir\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: E0106 14:50:21.940765 4744 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Jan 06 14:50:21 crc kubenswrapper[4744]: E0106 14:50:21.940837 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-metrics podName:8f0b0ccf-898d-4dbb-9be8-c37a780a7e85 nodeName:}" failed. No retries permitted until 2026-01-06 14:50:22.440816463 +0000 UTC m=+819.068282791 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-metrics") pod "collector-v2cmh" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85") : secret "collector-metrics" not found Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940679 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-trusted-ca\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.940962 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-syslog-receiver\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: E0106 14:50:21.941105 4744 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Jan 06 14:50:21 crc kubenswrapper[4744]: E0106 14:50:21.941185 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-syslog-receiver podName:8f0b0ccf-898d-4dbb-9be8-c37a780a7e85 nodeName:}" failed. No retries permitted until 2026-01-06 14:50:22.441149172 +0000 UTC m=+819.068615500 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-syslog-receiver") pod "collector-v2cmh" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85") : secret "collector-syslog-receiver" not found Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.941902 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config-openshift-service-cacrt\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.941916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-trusted-ca\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.941938 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.942340 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-entrypoint\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.948249 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-tmp\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.950111 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-token\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.972220 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-sa-token\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:21 crc kubenswrapper[4744]: I0106 14:50:21.972420 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g24gl\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-kube-api-access-g24gl\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.251292 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-v2cmh" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.260066 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-v2cmh" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346119 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-token\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346309 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346457 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-sa-token\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346565 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-trusted-ca\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346607 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g24gl\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-kube-api-access-g24gl\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346667 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-tmp\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346742 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-datadir\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346804 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config-openshift-service-cacrt\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346861 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-entrypoint\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346904 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-datadir" (OuterVolumeSpecName: "datadir") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.346998 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.347347 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config" (OuterVolumeSpecName: "config") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.347415 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.347557 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.347571 4744 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-datadir\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.347641 4744 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.347659 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.347675 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.349860 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-tmp" (OuterVolumeSpecName: "tmp") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.350237 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-sa-token" (OuterVolumeSpecName: "sa-token") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.351830 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-kube-api-access-g24gl" (OuterVolumeSpecName: "kube-api-access-g24gl") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "kube-api-access-g24gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.352622 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-token" (OuterVolumeSpecName: "collector-token") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.448674 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-syslog-receiver\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.449040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-metrics\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.449104 4744 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-token\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.449116 4744 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-sa-token\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.449127 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g24gl\" (UniqueName: \"kubernetes.io/projected/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-kube-api-access-g24gl\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.449137 4744 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-tmp\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.449145 4744 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-entrypoint\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.453518 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-metrics\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.458967 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-syslog-receiver\") pod \"collector-v2cmh\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " pod="openshift-logging/collector-v2cmh" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.550503 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-syslog-receiver\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.550874 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-metrics\") pod \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\" (UID: \"8f0b0ccf-898d-4dbb-9be8-c37a780a7e85\") " Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.553061 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.555011 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-metrics" (OuterVolumeSpecName: "metrics") pod "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" (UID: "8f0b0ccf-898d-4dbb-9be8-c37a780a7e85"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.652727 4744 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:22 crc kubenswrapper[4744]: I0106 14:50:22.652782 4744 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85-metrics\") on node \"crc\" DevicePath \"\"" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.262529 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-v2cmh" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.338903 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-v2cmh"] Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.349223 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-v2cmh"] Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.363376 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-95ctr"] Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.364579 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.370980 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-95ctr"] Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.371551 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.372388 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-7lhp8" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.372544 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.372684 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.372951 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.372953 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/636a1772-2f55-430d-a282-35d995eeb134-collector-token\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373284 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/636a1772-2f55-430d-a282-35d995eeb134-sa-token\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373311 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/636a1772-2f55-430d-a282-35d995eeb134-metrics\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373353 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-config-openshift-service-cacrt\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373397 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-entrypoint\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373421 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/636a1772-2f55-430d-a282-35d995eeb134-datadir\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373444 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/636a1772-2f55-430d-a282-35d995eeb134-tmp\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-trusted-ca\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373526 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmspc\" (UniqueName: \"kubernetes.io/projected/636a1772-2f55-430d-a282-35d995eeb134-kube-api-access-jmspc\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373555 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-config\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.373591 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/636a1772-2f55-430d-a282-35d995eeb134-collector-syslog-receiver\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.381254 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475395 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-config\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475454 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/636a1772-2f55-430d-a282-35d995eeb134-collector-syslog-receiver\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/636a1772-2f55-430d-a282-35d995eeb134-collector-token\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475534 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/636a1772-2f55-430d-a282-35d995eeb134-sa-token\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/636a1772-2f55-430d-a282-35d995eeb134-metrics\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475577 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-config-openshift-service-cacrt\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-entrypoint\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475628 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/636a1772-2f55-430d-a282-35d995eeb134-datadir\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475644 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/636a1772-2f55-430d-a282-35d995eeb134-tmp\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475664 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-trusted-ca\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.475682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmspc\" (UniqueName: \"kubernetes.io/projected/636a1772-2f55-430d-a282-35d995eeb134-kube-api-access-jmspc\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.476187 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/636a1772-2f55-430d-a282-35d995eeb134-datadir\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.476618 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-config-openshift-service-cacrt\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.476812 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-config\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.476939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-entrypoint\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.477678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/636a1772-2f55-430d-a282-35d995eeb134-trusted-ca\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.479147 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/636a1772-2f55-430d-a282-35d995eeb134-tmp\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.480748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/636a1772-2f55-430d-a282-35d995eeb134-collector-token\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.482541 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/636a1772-2f55-430d-a282-35d995eeb134-metrics\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.488398 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/636a1772-2f55-430d-a282-35d995eeb134-collector-syslog-receiver\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.490996 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmspc\" (UniqueName: \"kubernetes.io/projected/636a1772-2f55-430d-a282-35d995eeb134-kube-api-access-jmspc\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.492803 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/636a1772-2f55-430d-a282-35d995eeb134-sa-token\") pod \"collector-95ctr\" (UID: \"636a1772-2f55-430d-a282-35d995eeb134\") " pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.687879 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-95ctr" Jan 06 14:50:23 crc kubenswrapper[4744]: I0106 14:50:23.720536 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f0b0ccf-898d-4dbb-9be8-c37a780a7e85" path="/var/lib/kubelet/pods/8f0b0ccf-898d-4dbb-9be8-c37a780a7e85/volumes" Jan 06 14:50:24 crc kubenswrapper[4744]: I0106 14:50:24.111971 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-95ctr"] Jan 06 14:50:24 crc kubenswrapper[4744]: W0106 14:50:24.120305 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod636a1772_2f55_430d_a282_35d995eeb134.slice/crio-71c8b1bc61e4a94eb5d73167cc622d7bf3341408fe54495bb636b2fd6419cb47 WatchSource:0}: Error finding container 71c8b1bc61e4a94eb5d73167cc622d7bf3341408fe54495bb636b2fd6419cb47: Status 404 returned error can't find the container with id 71c8b1bc61e4a94eb5d73167cc622d7bf3341408fe54495bb636b2fd6419cb47 Jan 06 14:50:24 crc kubenswrapper[4744]: I0106 14:50:24.271271 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-95ctr" event={"ID":"636a1772-2f55-430d-a282-35d995eeb134","Type":"ContainerStarted","Data":"71c8b1bc61e4a94eb5d73167cc622d7bf3341408fe54495bb636b2fd6419cb47"} Jan 06 14:50:32 crc kubenswrapper[4744]: I0106 14:50:32.356096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-95ctr" event={"ID":"636a1772-2f55-430d-a282-35d995eeb134","Type":"ContainerStarted","Data":"8ec8c5e8345075201fe29369158bbe525a544dd8657ec53b114ad77953d8e310"} Jan 06 14:50:32 crc kubenswrapper[4744]: I0106 14:50:32.389591 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-95ctr" podStartSLOduration=2.223951619 podStartE2EDuration="9.389561747s" podCreationTimestamp="2026-01-06 14:50:23 +0000 UTC" firstStartedPulling="2026-01-06 14:50:24.123665816 +0000 UTC m=+820.751132134" lastFinishedPulling="2026-01-06 14:50:31.289275914 +0000 UTC m=+827.916742262" observedRunningTime="2026-01-06 14:50:32.379583635 +0000 UTC m=+829.007049963" watchObservedRunningTime="2026-01-06 14:50:32.389561747 +0000 UTC m=+829.017028105" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.280138 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2"] Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.282920 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.286621 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.297198 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2"] Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.384700 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.385123 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.385261 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qktw5\" (UniqueName: \"kubernetes.io/projected/7b10c0ca-7d3b-436c-be4e-80798af3e660-kube-api-access-qktw5\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.486129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qktw5\" (UniqueName: \"kubernetes.io/projected/7b10c0ca-7d3b-436c-be4e-80798af3e660-kube-api-access-qktw5\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.486226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.486284 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.486798 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.486839 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.507589 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qktw5\" (UniqueName: \"kubernetes.io/projected/7b10c0ca-7d3b-436c-be4e-80798af3e660-kube-api-access-qktw5\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:05 crc kubenswrapper[4744]: I0106 14:51:05.607557 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:06 crc kubenswrapper[4744]: I0106 14:51:06.118877 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2"] Jan 06 14:51:06 crc kubenswrapper[4744]: I0106 14:51:06.691525 4744 generic.go:334] "Generic (PLEG): container finished" podID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerID="8768def8b97b30c81d2e914c189efa0fc7ad86c3b5aebc218230d2baca5a93a4" exitCode=0 Jan 06 14:51:06 crc kubenswrapper[4744]: I0106 14:51:06.691577 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" event={"ID":"7b10c0ca-7d3b-436c-be4e-80798af3e660","Type":"ContainerDied","Data":"8768def8b97b30c81d2e914c189efa0fc7ad86c3b5aebc218230d2baca5a93a4"} Jan 06 14:51:06 crc kubenswrapper[4744]: I0106 14:51:06.691913 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" event={"ID":"7b10c0ca-7d3b-436c-be4e-80798af3e660","Type":"ContainerStarted","Data":"be293e422782fc75829c1bf36bc5f54c5adecfbc38df0f857aae7008cd578cc4"} Jan 06 14:51:08 crc kubenswrapper[4744]: I0106 14:51:08.716648 4744 generic.go:334] "Generic (PLEG): container finished" podID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerID="762968cfbe8b9fbf235976ce825955c4fb76ef2c5180bc6691e83bfdd007abde" exitCode=0 Jan 06 14:51:08 crc kubenswrapper[4744]: I0106 14:51:08.716896 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" event={"ID":"7b10c0ca-7d3b-436c-be4e-80798af3e660","Type":"ContainerDied","Data":"762968cfbe8b9fbf235976ce825955c4fb76ef2c5180bc6691e83bfdd007abde"} Jan 06 14:51:09 crc kubenswrapper[4744]: I0106 14:51:09.725796 4744 generic.go:334] "Generic (PLEG): container finished" podID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerID="54f979946fd6df93bceaef2d38426354eef3ae644983c395a3e916696ac75cc1" exitCode=0 Jan 06 14:51:09 crc kubenswrapper[4744]: I0106 14:51:09.725834 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" event={"ID":"7b10c0ca-7d3b-436c-be4e-80798af3e660","Type":"ContainerDied","Data":"54f979946fd6df93bceaef2d38426354eef3ae644983c395a3e916696ac75cc1"} Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.074077 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.195897 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-util\") pod \"7b10c0ca-7d3b-436c-be4e-80798af3e660\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.196048 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qktw5\" (UniqueName: \"kubernetes.io/projected/7b10c0ca-7d3b-436c-be4e-80798af3e660-kube-api-access-qktw5\") pod \"7b10c0ca-7d3b-436c-be4e-80798af3e660\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.196096 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-bundle\") pod \"7b10c0ca-7d3b-436c-be4e-80798af3e660\" (UID: \"7b10c0ca-7d3b-436c-be4e-80798af3e660\") " Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.197513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-bundle" (OuterVolumeSpecName: "bundle") pod "7b10c0ca-7d3b-436c-be4e-80798af3e660" (UID: "7b10c0ca-7d3b-436c-be4e-80798af3e660"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.204126 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b10c0ca-7d3b-436c-be4e-80798af3e660-kube-api-access-qktw5" (OuterVolumeSpecName: "kube-api-access-qktw5") pod "7b10c0ca-7d3b-436c-be4e-80798af3e660" (UID: "7b10c0ca-7d3b-436c-be4e-80798af3e660"). InnerVolumeSpecName "kube-api-access-qktw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.215172 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-util" (OuterVolumeSpecName: "util") pod "7b10c0ca-7d3b-436c-be4e-80798af3e660" (UID: "7b10c0ca-7d3b-436c-be4e-80798af3e660"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.298402 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-util\") on node \"crc\" DevicePath \"\"" Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.298450 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qktw5\" (UniqueName: \"kubernetes.io/projected/7b10c0ca-7d3b-436c-be4e-80798af3e660-kube-api-access-qktw5\") on node \"crc\" DevicePath \"\"" Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.298471 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7b10c0ca-7d3b-436c-be4e-80798af3e660-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.745104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" event={"ID":"7b10c0ca-7d3b-436c-be4e-80798af3e660","Type":"ContainerDied","Data":"be293e422782fc75829c1bf36bc5f54c5adecfbc38df0f857aae7008cd578cc4"} Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.745179 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be293e422782fc75829c1bf36bc5f54c5adecfbc38df0f857aae7008cd578cc4" Jan 06 14:51:11 crc kubenswrapper[4744]: I0106 14:51:11.745199 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2" Jan 06 14:51:14 crc kubenswrapper[4744]: I0106 14:51:14.426113 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:51:14 crc kubenswrapper[4744]: I0106 14:51:14.426445 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.240599 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-d8hmc"] Jan 06 14:51:18 crc kubenswrapper[4744]: E0106 14:51:18.241718 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerName="pull" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.241739 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerName="pull" Jan 06 14:51:18 crc kubenswrapper[4744]: E0106 14:51:18.241786 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerName="util" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.241797 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerName="util" Jan 06 14:51:18 crc kubenswrapper[4744]: E0106 14:51:18.241819 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerName="extract" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.241830 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerName="extract" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.242047 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b10c0ca-7d3b-436c-be4e-80798af3e660" containerName="extract" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.242847 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-d8hmc" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.247445 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.249025 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.249679 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-8f8b6" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.258063 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-d8hmc"] Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.426205 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hgg\" (UniqueName: \"kubernetes.io/projected/c6660ee8-e4a2-4898-b8fc-ea4354bfc879-kube-api-access-87hgg\") pod \"nmstate-operator-6769fb99d-d8hmc\" (UID: \"c6660ee8-e4a2-4898-b8fc-ea4354bfc879\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-d8hmc" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.527571 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hgg\" (UniqueName: \"kubernetes.io/projected/c6660ee8-e4a2-4898-b8fc-ea4354bfc879-kube-api-access-87hgg\") pod \"nmstate-operator-6769fb99d-d8hmc\" (UID: \"c6660ee8-e4a2-4898-b8fc-ea4354bfc879\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-d8hmc" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.553435 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hgg\" (UniqueName: \"kubernetes.io/projected/c6660ee8-e4a2-4898-b8fc-ea4354bfc879-kube-api-access-87hgg\") pod \"nmstate-operator-6769fb99d-d8hmc\" (UID: \"c6660ee8-e4a2-4898-b8fc-ea4354bfc879\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-d8hmc" Jan 06 14:51:18 crc kubenswrapper[4744]: I0106 14:51:18.568751 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-d8hmc" Jan 06 14:51:19 crc kubenswrapper[4744]: I0106 14:51:19.081922 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 14:51:19 crc kubenswrapper[4744]: I0106 14:51:19.081918 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-d8hmc"] Jan 06 14:51:19 crc kubenswrapper[4744]: I0106 14:51:19.808005 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-d8hmc" event={"ID":"c6660ee8-e4a2-4898-b8fc-ea4354bfc879","Type":"ContainerStarted","Data":"da258ad15e3f828f47b77d102c02429d2d0e1f8d216d9db1400bfa929d2f0395"} Jan 06 14:51:19 crc kubenswrapper[4744]: I0106 14:51:19.892353 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kwbx2"] Jan 06 14:51:19 crc kubenswrapper[4744]: I0106 14:51:19.894472 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:19 crc kubenswrapper[4744]: I0106 14:51:19.939071 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwbx2"] Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.052072 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zdlm\" (UniqueName: \"kubernetes.io/projected/1772fa34-1b0c-481a-80f3-1340b6626a69-kube-api-access-8zdlm\") pod \"redhat-marketplace-kwbx2\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.052135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-catalog-content\") pod \"redhat-marketplace-kwbx2\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.052208 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-utilities\") pod \"redhat-marketplace-kwbx2\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.153965 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zdlm\" (UniqueName: \"kubernetes.io/projected/1772fa34-1b0c-481a-80f3-1340b6626a69-kube-api-access-8zdlm\") pod \"redhat-marketplace-kwbx2\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.154019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-catalog-content\") pod \"redhat-marketplace-kwbx2\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.154064 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-utilities\") pod \"redhat-marketplace-kwbx2\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.154730 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-utilities\") pod \"redhat-marketplace-kwbx2\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.154946 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-catalog-content\") pod \"redhat-marketplace-kwbx2\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.180040 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zdlm\" (UniqueName: \"kubernetes.io/projected/1772fa34-1b0c-481a-80f3-1340b6626a69-kube-api-access-8zdlm\") pod \"redhat-marketplace-kwbx2\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.273114 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.705307 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwbx2"] Jan 06 14:51:20 crc kubenswrapper[4744]: I0106 14:51:20.818009 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwbx2" event={"ID":"1772fa34-1b0c-481a-80f3-1340b6626a69","Type":"ContainerStarted","Data":"5f0e269d8cb26c7213027d574132f6d6d0915f451cb2c18ae6a373e203fd32b4"} Jan 06 14:51:21 crc kubenswrapper[4744]: I0106 14:51:21.830711 4744 generic.go:334] "Generic (PLEG): container finished" podID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerID="a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df" exitCode=0 Jan 06 14:51:21 crc kubenswrapper[4744]: I0106 14:51:21.830904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwbx2" event={"ID":"1772fa34-1b0c-481a-80f3-1340b6626a69","Type":"ContainerDied","Data":"a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df"} Jan 06 14:51:22 crc kubenswrapper[4744]: I0106 14:51:22.842656 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-d8hmc" event={"ID":"c6660ee8-e4a2-4898-b8fc-ea4354bfc879","Type":"ContainerStarted","Data":"c158194812ee68c4cf3e9718bae007d18b404909c351ef8edd8cc0196e9a4fae"} Jan 06 14:51:22 crc kubenswrapper[4744]: I0106 14:51:22.845266 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwbx2" event={"ID":"1772fa34-1b0c-481a-80f3-1340b6626a69","Type":"ContainerStarted","Data":"82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569"} Jan 06 14:51:22 crc kubenswrapper[4744]: I0106 14:51:22.869659 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-6769fb99d-d8hmc" podStartSLOduration=2.090924405 podStartE2EDuration="4.869626093s" podCreationTimestamp="2026-01-06 14:51:18 +0000 UTC" firstStartedPulling="2026-01-06 14:51:19.081709088 +0000 UTC m=+875.709175406" lastFinishedPulling="2026-01-06 14:51:21.860410756 +0000 UTC m=+878.487877094" observedRunningTime="2026-01-06 14:51:22.863930644 +0000 UTC m=+879.491396982" watchObservedRunningTime="2026-01-06 14:51:22.869626093 +0000 UTC m=+879.497092461" Jan 06 14:51:23 crc kubenswrapper[4744]: I0106 14:51:23.856792 4744 generic.go:334] "Generic (PLEG): container finished" podID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerID="82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569" exitCode=0 Jan 06 14:51:23 crc kubenswrapper[4744]: I0106 14:51:23.856913 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwbx2" event={"ID":"1772fa34-1b0c-481a-80f3-1340b6626a69","Type":"ContainerDied","Data":"82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569"} Jan 06 14:51:24 crc kubenswrapper[4744]: I0106 14:51:24.866961 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwbx2" event={"ID":"1772fa34-1b0c-481a-80f3-1340b6626a69","Type":"ContainerStarted","Data":"1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3"} Jan 06 14:51:24 crc kubenswrapper[4744]: I0106 14:51:24.891292 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kwbx2" podStartSLOduration=3.429933901 podStartE2EDuration="5.891265452s" podCreationTimestamp="2026-01-06 14:51:19 +0000 UTC" firstStartedPulling="2026-01-06 14:51:21.853359101 +0000 UTC m=+878.480825419" lastFinishedPulling="2026-01-06 14:51:24.314690642 +0000 UTC m=+880.942156970" observedRunningTime="2026-01-06 14:51:24.888204301 +0000 UTC m=+881.515670659" watchObservedRunningTime="2026-01-06 14:51:24.891265452 +0000 UTC m=+881.518731800" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.693332 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw"] Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.694736 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.696736 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-nz6fp" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.707450 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw"] Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.723208 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-gkbrk"] Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.724237 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.743676 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q"] Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.744725 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.748805 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.760108 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q"] Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.840291 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8"] Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.841120 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.848525 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.848659 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.848723 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-4zplc" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.851977 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8"] Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.865912 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbzxg\" (UniqueName: \"kubernetes.io/projected/6488ead4-081d-4faa-b3f3-f76b713b4aa9-kube-api-access-rbzxg\") pod \"nmstate-metrics-7f7f7578db-fcpmw\" (UID: \"6488ead4-081d-4faa-b3f3-f76b713b4aa9\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.865950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e89278c2-e058-4ea3-823f-00438d7ece52-nmstate-lock\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.871927 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-v4x2q\" (UID: \"ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.872026 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjz97\" (UniqueName: \"kubernetes.io/projected/ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a-kube-api-access-xjz97\") pod \"nmstate-webhook-f8fb84555-v4x2q\" (UID: \"ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.872130 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e89278c2-e058-4ea3-823f-00438d7ece52-dbus-socket\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.872154 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e89278c2-e058-4ea3-823f-00438d7ece52-ovs-socket\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.872225 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9xtq\" (UniqueName: \"kubernetes.io/projected/e89278c2-e058-4ea3-823f-00438d7ece52-kube-api-access-d9xtq\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974105 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/13ffe872-9861-485f-95a0-deaf5d02c62f-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-9l9t8\" (UID: \"13ffe872-9861-485f-95a0-deaf5d02c62f\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974210 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e89278c2-e058-4ea3-823f-00438d7ece52-dbus-socket\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974229 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e89278c2-e058-4ea3-823f-00438d7ece52-ovs-socket\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9xtq\" (UniqueName: \"kubernetes.io/projected/e89278c2-e058-4ea3-823f-00438d7ece52-kube-api-access-d9xtq\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974304 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbzxg\" (UniqueName: \"kubernetes.io/projected/6488ead4-081d-4faa-b3f3-f76b713b4aa9-kube-api-access-rbzxg\") pod \"nmstate-metrics-7f7f7578db-fcpmw\" (UID: \"6488ead4-081d-4faa-b3f3-f76b713b4aa9\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e89278c2-e058-4ea3-823f-00438d7ece52-nmstate-lock\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974346 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsq75\" (UniqueName: \"kubernetes.io/projected/13ffe872-9861-485f-95a0-deaf5d02c62f-kube-api-access-dsq75\") pod \"nmstate-console-plugin-6ff7998486-9l9t8\" (UID: \"13ffe872-9861-485f-95a0-deaf5d02c62f\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974391 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-v4x2q\" (UID: \"ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974419 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjz97\" (UniqueName: \"kubernetes.io/projected/ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a-kube-api-access-xjz97\") pod \"nmstate-webhook-f8fb84555-v4x2q\" (UID: \"ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974440 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/13ffe872-9861-485f-95a0-deaf5d02c62f-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-9l9t8\" (UID: \"13ffe872-9861-485f-95a0-deaf5d02c62f\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974727 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/e89278c2-e058-4ea3-823f-00438d7ece52-dbus-socket\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.974812 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/e89278c2-e058-4ea3-823f-00438d7ece52-nmstate-lock\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: E0106 14:51:26.974821 4744 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Jan 06 14:51:26 crc kubenswrapper[4744]: E0106 14:51:26.974928 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a-tls-key-pair podName:ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a nodeName:}" failed. No retries permitted until 2026-01-06 14:51:27.474904154 +0000 UTC m=+884.102370472 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a-tls-key-pair") pod "nmstate-webhook-f8fb84555-v4x2q" (UID: "ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a") : secret "openshift-nmstate-webhook" not found Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.975107 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/e89278c2-e058-4ea3-823f-00438d7ece52-ovs-socket\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:26 crc kubenswrapper[4744]: I0106 14:51:26.997115 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjz97\" (UniqueName: \"kubernetes.io/projected/ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a-kube-api-access-xjz97\") pod \"nmstate-webhook-f8fb84555-v4x2q\" (UID: \"ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.008976 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbzxg\" (UniqueName: \"kubernetes.io/projected/6488ead4-081d-4faa-b3f3-f76b713b4aa9-kube-api-access-rbzxg\") pod \"nmstate-metrics-7f7f7578db-fcpmw\" (UID: \"6488ead4-081d-4faa-b3f3-f76b713b4aa9\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.020477 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.032332 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9xtq\" (UniqueName: \"kubernetes.io/projected/e89278c2-e058-4ea3-823f-00438d7ece52-kube-api-access-d9xtq\") pod \"nmstate-handler-gkbrk\" (UID: \"e89278c2-e058-4ea3-823f-00438d7ece52\") " pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.060602 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.077617 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/13ffe872-9861-485f-95a0-deaf5d02c62f-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-9l9t8\" (UID: \"13ffe872-9861-485f-95a0-deaf5d02c62f\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.077669 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/13ffe872-9861-485f-95a0-deaf5d02c62f-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-9l9t8\" (UID: \"13ffe872-9861-485f-95a0-deaf5d02c62f\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.077747 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsq75\" (UniqueName: \"kubernetes.io/projected/13ffe872-9861-485f-95a0-deaf5d02c62f-kube-api-access-dsq75\") pod \"nmstate-console-plugin-6ff7998486-9l9t8\" (UID: \"13ffe872-9861-485f-95a0-deaf5d02c62f\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.079434 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/13ffe872-9861-485f-95a0-deaf5d02c62f-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-9l9t8\" (UID: \"13ffe872-9861-485f-95a0-deaf5d02c62f\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.085772 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/13ffe872-9861-485f-95a0-deaf5d02c62f-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-9l9t8\" (UID: \"13ffe872-9861-485f-95a0-deaf5d02c62f\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.126193 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-857bd77bfc-rvs27"] Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.127116 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.130903 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsq75\" (UniqueName: \"kubernetes.io/projected/13ffe872-9861-485f-95a0-deaf5d02c62f-kube-api-access-dsq75\") pod \"nmstate-console-plugin-6ff7998486-9l9t8\" (UID: \"13ffe872-9861-485f-95a0-deaf5d02c62f\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.161976 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-857bd77bfc-rvs27"] Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.164941 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.281215 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-serving-cert\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.281439 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-service-ca\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.281508 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-config\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.281543 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-trusted-ca-bundle\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.282307 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-oauth-config\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.282590 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-oauth-serving-cert\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.282763 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-src2w\" (UniqueName: \"kubernetes.io/projected/4b070bab-3e2c-4579-8d78-92ad1f973e62-kube-api-access-src2w\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.384542 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-trusted-ca-bundle\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.384595 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-oauth-config\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.384632 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-oauth-serving-cert\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.384663 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-src2w\" (UniqueName: \"kubernetes.io/projected/4b070bab-3e2c-4579-8d78-92ad1f973e62-kube-api-access-src2w\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.384702 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-serving-cert\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.384733 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-service-ca\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.384772 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-config\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.385601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-trusted-ca-bundle\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.385624 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-service-ca\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.385604 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-config\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.385667 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-oauth-serving-cert\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.389700 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-oauth-config\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.392757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-serving-cert\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.399300 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-src2w\" (UniqueName: \"kubernetes.io/projected/4b070bab-3e2c-4579-8d78-92ad1f973e62-kube-api-access-src2w\") pod \"console-857bd77bfc-rvs27\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.455860 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.487431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-v4x2q\" (UID: \"ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.490458 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-v4x2q\" (UID: \"ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.510972 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8"] Jan 06 14:51:27 crc kubenswrapper[4744]: W0106 14:51:27.514510 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13ffe872_9861_485f_95a0_deaf5d02c62f.slice/crio-0906db0e31ab478c1ae50846d70e7aeb043dc58dd04bccd0ff2ca1546ffb61b1 WatchSource:0}: Error finding container 0906db0e31ab478c1ae50846d70e7aeb043dc58dd04bccd0ff2ca1546ffb61b1: Status 404 returned error can't find the container with id 0906db0e31ab478c1ae50846d70e7aeb043dc58dd04bccd0ff2ca1546ffb61b1 Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.540521 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw"] Jan 06 14:51:27 crc kubenswrapper[4744]: W0106 14:51:27.556858 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6488ead4_081d_4faa_b3f3_f76b713b4aa9.slice/crio-0a393fb9045607b9264bacae1c5521d8f45425d3b329d2c225b54032bf56a50d WatchSource:0}: Error finding container 0a393fb9045607b9264bacae1c5521d8f45425d3b329d2c225b54032bf56a50d: Status 404 returned error can't find the container with id 0a393fb9045607b9264bacae1c5521d8f45425d3b329d2c225b54032bf56a50d Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.665846 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.882400 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-857bd77bfc-rvs27"] Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.909478 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-857bd77bfc-rvs27" event={"ID":"4b070bab-3e2c-4579-8d78-92ad1f973e62","Type":"ContainerStarted","Data":"74ffafa99f41e0ef114ec6d3df28c1e9201cda09c4332b19f6a5748bd04b96fd"} Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.913257 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" event={"ID":"13ffe872-9861-485f-95a0-deaf5d02c62f","Type":"ContainerStarted","Data":"0906db0e31ab478c1ae50846d70e7aeb043dc58dd04bccd0ff2ca1546ffb61b1"} Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.916928 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gkbrk" event={"ID":"e89278c2-e058-4ea3-823f-00438d7ece52","Type":"ContainerStarted","Data":"104d7cb0fe3e10468a3a3ba82e8315ba9f1c210a21b93568f52f44cfb7e227d0"} Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.924227 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw" event={"ID":"6488ead4-081d-4faa-b3f3-f76b713b4aa9","Type":"ContainerStarted","Data":"0a393fb9045607b9264bacae1c5521d8f45425d3b329d2c225b54032bf56a50d"} Jan 06 14:51:27 crc kubenswrapper[4744]: I0106 14:51:27.931320 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q"] Jan 06 14:51:28 crc kubenswrapper[4744]: I0106 14:51:28.949972 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-857bd77bfc-rvs27" event={"ID":"4b070bab-3e2c-4579-8d78-92ad1f973e62","Type":"ContainerStarted","Data":"0f1a7c93518967526038315d0f2e18176895b43b2cfeeb421285a669ffb57ef8"} Jan 06 14:51:28 crc kubenswrapper[4744]: I0106 14:51:28.951349 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" event={"ID":"ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a","Type":"ContainerStarted","Data":"43814e445e7134ee43f6ae10f0c54b184a85f3e9de4b9074fe0d02afbf24263c"} Jan 06 14:51:28 crc kubenswrapper[4744]: I0106 14:51:28.970668 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-857bd77bfc-rvs27" podStartSLOduration=1.970646143 podStartE2EDuration="1.970646143s" podCreationTimestamp="2026-01-06 14:51:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:51:28.964698307 +0000 UTC m=+885.592164625" watchObservedRunningTime="2026-01-06 14:51:28.970646143 +0000 UTC m=+885.598112461" Jan 06 14:51:30 crc kubenswrapper[4744]: I0106 14:51:30.273822 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:30 crc kubenswrapper[4744]: I0106 14:51:30.273934 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:30 crc kubenswrapper[4744]: I0106 14:51:30.316321 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:31 crc kubenswrapper[4744]: I0106 14:51:31.035628 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:31 crc kubenswrapper[4744]: I0106 14:51:31.997184 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" event={"ID":"ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a","Type":"ContainerStarted","Data":"4d784e2544e8e5d4b243510a06980ba09a011414dd0c39a336773b2d05428f5e"} Jan 06 14:51:31 crc kubenswrapper[4744]: I0106 14:51:31.997532 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:51:32 crc kubenswrapper[4744]: I0106 14:51:32.003675 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" event={"ID":"13ffe872-9861-485f-95a0-deaf5d02c62f","Type":"ContainerStarted","Data":"0c2c38c02c333d2ab5ad6e849c12a16726600c18438764f3bca277fdeb243855"} Jan 06 14:51:32 crc kubenswrapper[4744]: I0106 14:51:32.022390 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" podStartSLOduration=2.260093404 podStartE2EDuration="6.022375127s" podCreationTimestamp="2026-01-06 14:51:26 +0000 UTC" firstStartedPulling="2026-01-06 14:51:27.962044933 +0000 UTC m=+884.589511251" lastFinishedPulling="2026-01-06 14:51:31.724326646 +0000 UTC m=+888.351792974" observedRunningTime="2026-01-06 14:51:32.015500766 +0000 UTC m=+888.642967084" watchObservedRunningTime="2026-01-06 14:51:32.022375127 +0000 UTC m=+888.649841435" Jan 06 14:51:32 crc kubenswrapper[4744]: I0106 14:51:32.038464 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-9l9t8" podStartSLOduration=1.833650859 podStartE2EDuration="6.038449338s" podCreationTimestamp="2026-01-06 14:51:26 +0000 UTC" firstStartedPulling="2026-01-06 14:51:27.516792965 +0000 UTC m=+884.144259283" lastFinishedPulling="2026-01-06 14:51:31.721591444 +0000 UTC m=+888.349057762" observedRunningTime="2026-01-06 14:51:32.034018132 +0000 UTC m=+888.661484450" watchObservedRunningTime="2026-01-06 14:51:32.038449338 +0000 UTC m=+888.665915656" Jan 06 14:51:32 crc kubenswrapper[4744]: I0106 14:51:32.080497 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwbx2"] Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.016770 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw" event={"ID":"6488ead4-081d-4faa-b3f3-f76b713b4aa9","Type":"ContainerStarted","Data":"52a094cb24be00f4cf77ee7311a523c26002997526b194aed724509b91b04622"} Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.018633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gkbrk" event={"ID":"e89278c2-e058-4ea3-823f-00438d7ece52","Type":"ContainerStarted","Data":"dbc0f7b70605bbb28c6edf0e1b5ab6829954f5a6ae048acc104454607210c01a"} Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.018932 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kwbx2" podUID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerName="registry-server" containerID="cri-o://1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3" gracePeriod=2 Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.019556 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.044599 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-gkbrk" podStartSLOduration=2.448069528 podStartE2EDuration="7.044579893s" podCreationTimestamp="2026-01-06 14:51:26 +0000 UTC" firstStartedPulling="2026-01-06 14:51:27.148878703 +0000 UTC m=+883.776345021" lastFinishedPulling="2026-01-06 14:51:31.745389068 +0000 UTC m=+888.372855386" observedRunningTime="2026-01-06 14:51:33.038044872 +0000 UTC m=+889.665511220" watchObservedRunningTime="2026-01-06 14:51:33.044579893 +0000 UTC m=+889.672046231" Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.483011 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.505352 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-catalog-content\") pod \"1772fa34-1b0c-481a-80f3-1340b6626a69\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.505427 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-utilities\") pod \"1772fa34-1b0c-481a-80f3-1340b6626a69\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.505472 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zdlm\" (UniqueName: \"kubernetes.io/projected/1772fa34-1b0c-481a-80f3-1340b6626a69-kube-api-access-8zdlm\") pod \"1772fa34-1b0c-481a-80f3-1340b6626a69\" (UID: \"1772fa34-1b0c-481a-80f3-1340b6626a69\") " Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.506442 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-utilities" (OuterVolumeSpecName: "utilities") pod "1772fa34-1b0c-481a-80f3-1340b6626a69" (UID: "1772fa34-1b0c-481a-80f3-1340b6626a69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.518102 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1772fa34-1b0c-481a-80f3-1340b6626a69-kube-api-access-8zdlm" (OuterVolumeSpecName: "kube-api-access-8zdlm") pod "1772fa34-1b0c-481a-80f3-1340b6626a69" (UID: "1772fa34-1b0c-481a-80f3-1340b6626a69"). InnerVolumeSpecName "kube-api-access-8zdlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.539109 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1772fa34-1b0c-481a-80f3-1340b6626a69" (UID: "1772fa34-1b0c-481a-80f3-1340b6626a69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.612666 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.612734 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zdlm\" (UniqueName: \"kubernetes.io/projected/1772fa34-1b0c-481a-80f3-1340b6626a69-kube-api-access-8zdlm\") on node \"crc\" DevicePath \"\"" Jan 06 14:51:33 crc kubenswrapper[4744]: I0106 14:51:33.612762 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1772fa34-1b0c-481a-80f3-1340b6626a69-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.034951 4744 generic.go:334] "Generic (PLEG): container finished" podID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerID="1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3" exitCode=0 Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.035058 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kwbx2" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.035122 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwbx2" event={"ID":"1772fa34-1b0c-481a-80f3-1340b6626a69","Type":"ContainerDied","Data":"1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3"} Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.035202 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kwbx2" event={"ID":"1772fa34-1b0c-481a-80f3-1340b6626a69","Type":"ContainerDied","Data":"5f0e269d8cb26c7213027d574132f6d6d0915f451cb2c18ae6a373e203fd32b4"} Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.035234 4744 scope.go:117] "RemoveContainer" containerID="1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.101497 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwbx2"] Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.101554 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kwbx2"] Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.749829 4744 scope.go:117] "RemoveContainer" containerID="82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.774437 4744 scope.go:117] "RemoveContainer" containerID="a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.836588 4744 scope.go:117] "RemoveContainer" containerID="1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3" Jan 06 14:51:34 crc kubenswrapper[4744]: E0106 14:51:34.837300 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3\": container with ID starting with 1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3 not found: ID does not exist" containerID="1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.837340 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3"} err="failed to get container status \"1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3\": rpc error: code = NotFound desc = could not find container \"1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3\": container with ID starting with 1c0992d498ee79dbf369657a78a86ea88bf0f04d55d2563dfb1bce5506657eb3 not found: ID does not exist" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.837365 4744 scope.go:117] "RemoveContainer" containerID="82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569" Jan 06 14:51:34 crc kubenswrapper[4744]: E0106 14:51:34.838452 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569\": container with ID starting with 82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569 not found: ID does not exist" containerID="82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.838495 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569"} err="failed to get container status \"82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569\": rpc error: code = NotFound desc = could not find container \"82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569\": container with ID starting with 82a7cc8b1bd1025c3cf6b279a8d29091b713904542e7370ce4fc4a8a649ea569 not found: ID does not exist" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.838523 4744 scope.go:117] "RemoveContainer" containerID="a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df" Jan 06 14:51:34 crc kubenswrapper[4744]: E0106 14:51:34.838894 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df\": container with ID starting with a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df not found: ID does not exist" containerID="a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df" Jan 06 14:51:34 crc kubenswrapper[4744]: I0106 14:51:34.838930 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df"} err="failed to get container status \"a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df\": rpc error: code = NotFound desc = could not find container \"a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df\": container with ID starting with a99dc1fac17ee0c55f4b18aa0ed353c15d708b7cb68bcbb98a56c02a7ddab3df not found: ID does not exist" Jan 06 14:51:35 crc kubenswrapper[4744]: I0106 14:51:35.719730 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1772fa34-1b0c-481a-80f3-1340b6626a69" path="/var/lib/kubelet/pods/1772fa34-1b0c-481a-80f3-1340b6626a69/volumes" Jan 06 14:51:36 crc kubenswrapper[4744]: I0106 14:51:36.053033 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw" event={"ID":"6488ead4-081d-4faa-b3f3-f76b713b4aa9","Type":"ContainerStarted","Data":"07a8c5e8271390643c254a99fd790ca0308d5367f0142037f8a5cb53e094bb5c"} Jan 06 14:51:36 crc kubenswrapper[4744]: I0106 14:51:36.073830 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-fcpmw" podStartSLOduration=2.703183005 podStartE2EDuration="10.073807096s" podCreationTimestamp="2026-01-06 14:51:26 +0000 UTC" firstStartedPulling="2026-01-06 14:51:27.560628804 +0000 UTC m=+884.188095122" lastFinishedPulling="2026-01-06 14:51:34.931252895 +0000 UTC m=+891.558719213" observedRunningTime="2026-01-06 14:51:36.071363222 +0000 UTC m=+892.698829580" watchObservedRunningTime="2026-01-06 14:51:36.073807096 +0000 UTC m=+892.701273424" Jan 06 14:51:37 crc kubenswrapper[4744]: I0106 14:51:37.092959 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-gkbrk" Jan 06 14:51:37 crc kubenswrapper[4744]: I0106 14:51:37.456065 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:37 crc kubenswrapper[4744]: I0106 14:51:37.456609 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:37 crc kubenswrapper[4744]: I0106 14:51:37.461821 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:38 crc kubenswrapper[4744]: I0106 14:51:38.078790 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:51:38 crc kubenswrapper[4744]: I0106 14:51:38.234536 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5dddd7d494-hhcc5"] Jan 06 14:51:44 crc kubenswrapper[4744]: I0106 14:51:44.424251 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:51:44 crc kubenswrapper[4744]: I0106 14:51:44.426884 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:51:47 crc kubenswrapper[4744]: I0106 14:51:47.673728 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-f8fb84555-v4x2q" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.276064 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-5dddd7d494-hhcc5" podUID="cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" containerName="console" containerID="cri-o://6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974" gracePeriod=15 Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.703493 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5dddd7d494-hhcc5_cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25/console/0.log" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.703798 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.771464 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-service-ca\") pod \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.773465 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-service-ca" (OuterVolumeSpecName: "service-ca") pod "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" (UID: "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.873299 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-oauth-serving-cert\") pod \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.873344 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-trusted-ca-bundle\") pod \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.873434 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsqln\" (UniqueName: \"kubernetes.io/projected/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-kube-api-access-hsqln\") pod \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.873457 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-config\") pod \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.873614 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-serving-cert\") pod \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.873645 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-oauth-config\") pod \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\" (UID: \"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25\") " Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.874287 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" (UID: "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.874309 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" (UID: "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.874422 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-config" (OuterVolumeSpecName: "console-config") pod "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" (UID: "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.875354 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.875372 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-service-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.875383 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.875392 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.880378 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-kube-api-access-hsqln" (OuterVolumeSpecName: "kube-api-access-hsqln") pod "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" (UID: "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25"). InnerVolumeSpecName "kube-api-access-hsqln". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.880812 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" (UID: "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.882638 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" (UID: "cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.976807 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.976845 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:03 crc kubenswrapper[4744]: I0106 14:52:03.976858 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsqln\" (UniqueName: \"kubernetes.io/projected/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25-kube-api-access-hsqln\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.362656 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5dddd7d494-hhcc5_cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25/console/0.log" Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.363152 4744 generic.go:334] "Generic (PLEG): container finished" podID="cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" containerID="6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974" exitCode=2 Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.363204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dddd7d494-hhcc5" event={"ID":"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25","Type":"ContainerDied","Data":"6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974"} Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.363237 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5dddd7d494-hhcc5" event={"ID":"cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25","Type":"ContainerDied","Data":"b6a278793ccc099f8e11894e806a9c3d007fceb8db60793e9f0ab1163f8f23ef"} Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.363258 4744 scope.go:117] "RemoveContainer" containerID="6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974" Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.363281 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5dddd7d494-hhcc5" Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.401618 4744 scope.go:117] "RemoveContainer" containerID="6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974" Jan 06 14:52:04 crc kubenswrapper[4744]: E0106 14:52:04.402106 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974\": container with ID starting with 6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974 not found: ID does not exist" containerID="6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974" Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.402187 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974"} err="failed to get container status \"6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974\": rpc error: code = NotFound desc = could not find container \"6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974\": container with ID starting with 6883d5fa0fe1c3d40f875da5b4fd03a59fe826ac5aadb9902085f957de79f974 not found: ID does not exist" Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.403008 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5dddd7d494-hhcc5"] Jan 06 14:52:04 crc kubenswrapper[4744]: I0106 14:52:04.411450 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5dddd7d494-hhcc5"] Jan 06 14:52:05 crc kubenswrapper[4744]: I0106 14:52:05.724272 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" path="/var/lib/kubelet/pods/cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25/volumes" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.760028 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt"] Jan 06 14:52:06 crc kubenswrapper[4744]: E0106 14:52:06.760726 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerName="extract-utilities" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.760743 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerName="extract-utilities" Jan 06 14:52:06 crc kubenswrapper[4744]: E0106 14:52:06.760760 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" containerName="console" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.760769 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" containerName="console" Jan 06 14:52:06 crc kubenswrapper[4744]: E0106 14:52:06.760797 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerName="registry-server" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.760806 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerName="registry-server" Jan 06 14:52:06 crc kubenswrapper[4744]: E0106 14:52:06.760824 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerName="extract-content" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.760832 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerName="extract-content" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.761021 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf6de0ec-f51e-4ac1-bb55-ea20d2bf8e25" containerName="console" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.761038 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1772fa34-1b0c-481a-80f3-1340b6626a69" containerName="registry-server" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.762479 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.765286 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.776132 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt"] Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.830432 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.830541 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt79j\" (UniqueName: \"kubernetes.io/projected/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-kube-api-access-lt79j\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.830607 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.932621 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt79j\" (UniqueName: \"kubernetes.io/projected/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-kube-api-access-lt79j\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.932746 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.932803 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.933523 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.933619 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:06 crc kubenswrapper[4744]: I0106 14:52:06.964968 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt79j\" (UniqueName: \"kubernetes.io/projected/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-kube-api-access-lt79j\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:07 crc kubenswrapper[4744]: I0106 14:52:07.088350 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:07 crc kubenswrapper[4744]: I0106 14:52:07.527098 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt"] Jan 06 14:52:08 crc kubenswrapper[4744]: I0106 14:52:08.394992 4744 generic.go:334] "Generic (PLEG): container finished" podID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerID="d80dd2ec4e054bb45163c3b83ace3a30566a9f349b3f7af74868c9d6c77b5956" exitCode=0 Jan 06 14:52:08 crc kubenswrapper[4744]: I0106 14:52:08.395065 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" event={"ID":"e1fc4c9c-6c58-420a-9132-1c9926cd00eb","Type":"ContainerDied","Data":"d80dd2ec4e054bb45163c3b83ace3a30566a9f349b3f7af74868c9d6c77b5956"} Jan 06 14:52:08 crc kubenswrapper[4744]: I0106 14:52:08.395390 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" event={"ID":"e1fc4c9c-6c58-420a-9132-1c9926cd00eb","Type":"ContainerStarted","Data":"9c8a13cd7413fc9ea5643d126ea9dda9bf282a505054a10b1458dc82cd80e4ef"} Jan 06 14:52:10 crc kubenswrapper[4744]: I0106 14:52:10.411668 4744 generic.go:334] "Generic (PLEG): container finished" podID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerID="5f1cd6ee4ff56c9d0894453c64fb45eaf2d56ac20f7fbae4f304aabe53191785" exitCode=0 Jan 06 14:52:10 crc kubenswrapper[4744]: I0106 14:52:10.411735 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" event={"ID":"e1fc4c9c-6c58-420a-9132-1c9926cd00eb","Type":"ContainerDied","Data":"5f1cd6ee4ff56c9d0894453c64fb45eaf2d56ac20f7fbae4f304aabe53191785"} Jan 06 14:52:11 crc kubenswrapper[4744]: I0106 14:52:11.420196 4744 generic.go:334] "Generic (PLEG): container finished" podID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerID="0022aa3ada1ec79ced3689696c02790cfa222b6a88e0eba7869b96a3da23a866" exitCode=0 Jan 06 14:52:11 crc kubenswrapper[4744]: I0106 14:52:11.420234 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" event={"ID":"e1fc4c9c-6c58-420a-9132-1c9926cd00eb","Type":"ContainerDied","Data":"0022aa3ada1ec79ced3689696c02790cfa222b6a88e0eba7869b96a3da23a866"} Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.711832 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.739144 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-bundle\") pod \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.739260 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt79j\" (UniqueName: \"kubernetes.io/projected/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-kube-api-access-lt79j\") pod \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.739312 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-util\") pod \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\" (UID: \"e1fc4c9c-6c58-420a-9132-1c9926cd00eb\") " Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.740303 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-bundle" (OuterVolumeSpecName: "bundle") pod "e1fc4c9c-6c58-420a-9132-1c9926cd00eb" (UID: "e1fc4c9c-6c58-420a-9132-1c9926cd00eb"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.750360 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-kube-api-access-lt79j" (OuterVolumeSpecName: "kube-api-access-lt79j") pod "e1fc4c9c-6c58-420a-9132-1c9926cd00eb" (UID: "e1fc4c9c-6c58-420a-9132-1c9926cd00eb"). InnerVolumeSpecName "kube-api-access-lt79j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.753295 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-util" (OuterVolumeSpecName: "util") pod "e1fc4c9c-6c58-420a-9132-1c9926cd00eb" (UID: "e1fc4c9c-6c58-420a-9132-1c9926cd00eb"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.841541 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.841574 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt79j\" (UniqueName: \"kubernetes.io/projected/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-kube-api-access-lt79j\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:12 crc kubenswrapper[4744]: I0106 14:52:12.841586 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e1fc4c9c-6c58-420a-9132-1c9926cd00eb-util\") on node \"crc\" DevicePath \"\"" Jan 06 14:52:13 crc kubenswrapper[4744]: I0106 14:52:13.436966 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" event={"ID":"e1fc4c9c-6c58-420a-9132-1c9926cd00eb","Type":"ContainerDied","Data":"9c8a13cd7413fc9ea5643d126ea9dda9bf282a505054a10b1458dc82cd80e4ef"} Jan 06 14:52:13 crc kubenswrapper[4744]: I0106 14:52:13.437012 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c8a13cd7413fc9ea5643d126ea9dda9bf282a505054a10b1458dc82cd80e4ef" Jan 06 14:52:13 crc kubenswrapper[4744]: I0106 14:52:13.437057 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt" Jan 06 14:52:14 crc kubenswrapper[4744]: I0106 14:52:14.423603 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:52:14 crc kubenswrapper[4744]: I0106 14:52:14.423652 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:52:14 crc kubenswrapper[4744]: I0106 14:52:14.423693 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:52:14 crc kubenswrapper[4744]: I0106 14:52:14.424311 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b7fdf0a873dd4feb170b2380a6eb8f3f910485e646fb343c689eab9c501fa171"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 14:52:14 crc kubenswrapper[4744]: I0106 14:52:14.424362 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://b7fdf0a873dd4feb170b2380a6eb8f3f910485e646fb343c689eab9c501fa171" gracePeriod=600 Jan 06 14:52:15 crc kubenswrapper[4744]: I0106 14:52:15.456685 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="b7fdf0a873dd4feb170b2380a6eb8f3f910485e646fb343c689eab9c501fa171" exitCode=0 Jan 06 14:52:15 crc kubenswrapper[4744]: I0106 14:52:15.456744 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"b7fdf0a873dd4feb170b2380a6eb8f3f910485e646fb343c689eab9c501fa171"} Jan 06 14:52:15 crc kubenswrapper[4744]: I0106 14:52:15.457125 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"7b95bf86d9d16dc50821869dcd40536e357bac00adf9429e476ea048ffe56203"} Jan 06 14:52:15 crc kubenswrapper[4744]: I0106 14:52:15.457147 4744 scope.go:117] "RemoveContainer" containerID="098e8945831f6c03a5214e0dd27553b61625d63e6339c0ac5bf8762f7f51444d" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.492602 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-744888559c-nspzf"] Jan 06 14:52:22 crc kubenswrapper[4744]: E0106 14:52:22.493504 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerName="pull" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.493519 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerName="pull" Jan 06 14:52:22 crc kubenswrapper[4744]: E0106 14:52:22.493541 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerName="util" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.493549 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerName="util" Jan 06 14:52:22 crc kubenswrapper[4744]: E0106 14:52:22.493568 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerName="extract" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.493575 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerName="extract" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.493767 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1fc4c9c-6c58-420a-9132-1c9926cd00eb" containerName="extract" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.494515 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.497243 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-5kkr7" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.497686 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.497814 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.497928 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.502589 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.516273 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-744888559c-nspzf"] Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.604951 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvxk9\" (UniqueName: \"kubernetes.io/projected/aadee3b1-2fa0-497f-b0cf-6e84bbf73d52-kube-api-access-vvxk9\") pod \"metallb-operator-controller-manager-744888559c-nspzf\" (UID: \"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52\") " pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.605008 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aadee3b1-2fa0-497f-b0cf-6e84bbf73d52-apiservice-cert\") pod \"metallb-operator-controller-manager-744888559c-nspzf\" (UID: \"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52\") " pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.605331 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aadee3b1-2fa0-497f-b0cf-6e84bbf73d52-webhook-cert\") pod \"metallb-operator-controller-manager-744888559c-nspzf\" (UID: \"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52\") " pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.707225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aadee3b1-2fa0-497f-b0cf-6e84bbf73d52-webhook-cert\") pod \"metallb-operator-controller-manager-744888559c-nspzf\" (UID: \"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52\") " pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.707331 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvxk9\" (UniqueName: \"kubernetes.io/projected/aadee3b1-2fa0-497f-b0cf-6e84bbf73d52-kube-api-access-vvxk9\") pod \"metallb-operator-controller-manager-744888559c-nspzf\" (UID: \"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52\") " pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.707364 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aadee3b1-2fa0-497f-b0cf-6e84bbf73d52-apiservice-cert\") pod \"metallb-operator-controller-manager-744888559c-nspzf\" (UID: \"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52\") " pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.714409 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/aadee3b1-2fa0-497f-b0cf-6e84bbf73d52-apiservice-cert\") pod \"metallb-operator-controller-manager-744888559c-nspzf\" (UID: \"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52\") " pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.714419 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/aadee3b1-2fa0-497f-b0cf-6e84bbf73d52-webhook-cert\") pod \"metallb-operator-controller-manager-744888559c-nspzf\" (UID: \"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52\") " pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.731319 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvxk9\" (UniqueName: \"kubernetes.io/projected/aadee3b1-2fa0-497f-b0cf-6e84bbf73d52-kube-api-access-vvxk9\") pod \"metallb-operator-controller-manager-744888559c-nspzf\" (UID: \"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52\") " pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.811194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.855421 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz"] Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.856537 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.859276 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.860225 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-jhtpw" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.862340 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.875501 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz"] Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.910091 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4447713-2058-42f3-9b7f-0225267625e3-webhook-cert\") pod \"metallb-operator-webhook-server-74784fd878-9zcxz\" (UID: \"c4447713-2058-42f3-9b7f-0225267625e3\") " pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.910253 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4447713-2058-42f3-9b7f-0225267625e3-apiservice-cert\") pod \"metallb-operator-webhook-server-74784fd878-9zcxz\" (UID: \"c4447713-2058-42f3-9b7f-0225267625e3\") " pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:22 crc kubenswrapper[4744]: I0106 14:52:22.910330 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scjqz\" (UniqueName: \"kubernetes.io/projected/c4447713-2058-42f3-9b7f-0225267625e3-kube-api-access-scjqz\") pod \"metallb-operator-webhook-server-74784fd878-9zcxz\" (UID: \"c4447713-2058-42f3-9b7f-0225267625e3\") " pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.013075 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4447713-2058-42f3-9b7f-0225267625e3-webhook-cert\") pod \"metallb-operator-webhook-server-74784fd878-9zcxz\" (UID: \"c4447713-2058-42f3-9b7f-0225267625e3\") " pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.013171 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4447713-2058-42f3-9b7f-0225267625e3-apiservice-cert\") pod \"metallb-operator-webhook-server-74784fd878-9zcxz\" (UID: \"c4447713-2058-42f3-9b7f-0225267625e3\") " pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.013201 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scjqz\" (UniqueName: \"kubernetes.io/projected/c4447713-2058-42f3-9b7f-0225267625e3-kube-api-access-scjqz\") pod \"metallb-operator-webhook-server-74784fd878-9zcxz\" (UID: \"c4447713-2058-42f3-9b7f-0225267625e3\") " pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.018677 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c4447713-2058-42f3-9b7f-0225267625e3-webhook-cert\") pod \"metallb-operator-webhook-server-74784fd878-9zcxz\" (UID: \"c4447713-2058-42f3-9b7f-0225267625e3\") " pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.038834 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scjqz\" (UniqueName: \"kubernetes.io/projected/c4447713-2058-42f3-9b7f-0225267625e3-kube-api-access-scjqz\") pod \"metallb-operator-webhook-server-74784fd878-9zcxz\" (UID: \"c4447713-2058-42f3-9b7f-0225267625e3\") " pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.053456 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c4447713-2058-42f3-9b7f-0225267625e3-apiservice-cert\") pod \"metallb-operator-webhook-server-74784fd878-9zcxz\" (UID: \"c4447713-2058-42f3-9b7f-0225267625e3\") " pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.203695 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.327663 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-744888559c-nspzf"] Jan 06 14:52:23 crc kubenswrapper[4744]: W0106 14:52:23.347401 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaadee3b1_2fa0_497f_b0cf_6e84bbf73d52.slice/crio-ff7d42c212f912b4fabccc8620473ddf37317cd96e8d5ef595e9920de849134b WatchSource:0}: Error finding container ff7d42c212f912b4fabccc8620473ddf37317cd96e8d5ef595e9920de849134b: Status 404 returned error can't find the container with id ff7d42c212f912b4fabccc8620473ddf37317cd96e8d5ef595e9920de849134b Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.521515 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" event={"ID":"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52","Type":"ContainerStarted","Data":"ff7d42c212f912b4fabccc8620473ddf37317cd96e8d5ef595e9920de849134b"} Jan 06 14:52:23 crc kubenswrapper[4744]: I0106 14:52:23.664442 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz"] Jan 06 14:52:23 crc kubenswrapper[4744]: W0106 14:52:23.664659 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4447713_2058_42f3_9b7f_0225267625e3.slice/crio-e153928ae597c15137ad50524ebc622d75e78f1e9b3521123722fe5616f22e80 WatchSource:0}: Error finding container e153928ae597c15137ad50524ebc622d75e78f1e9b3521123722fe5616f22e80: Status 404 returned error can't find the container with id e153928ae597c15137ad50524ebc622d75e78f1e9b3521123722fe5616f22e80 Jan 06 14:52:24 crc kubenswrapper[4744]: I0106 14:52:24.551688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" event={"ID":"c4447713-2058-42f3-9b7f-0225267625e3","Type":"ContainerStarted","Data":"e153928ae597c15137ad50524ebc622d75e78f1e9b3521123722fe5616f22e80"} Jan 06 14:52:27 crc kubenswrapper[4744]: I0106 14:52:27.584760 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" event={"ID":"aadee3b1-2fa0-497f-b0cf-6e84bbf73d52","Type":"ContainerStarted","Data":"d2fd1b491905b0e7069b7a9eacd4caf0a4e776424c36c457022f35460ac37542"} Jan 06 14:52:27 crc kubenswrapper[4744]: I0106 14:52:27.585501 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:52:27 crc kubenswrapper[4744]: I0106 14:52:27.609233 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" podStartSLOduration=2.219591687 podStartE2EDuration="5.609218873s" podCreationTimestamp="2026-01-06 14:52:22 +0000 UTC" firstStartedPulling="2026-01-06 14:52:23.353764237 +0000 UTC m=+939.981230555" lastFinishedPulling="2026-01-06 14:52:26.743391423 +0000 UTC m=+943.370857741" observedRunningTime="2026-01-06 14:52:27.605999438 +0000 UTC m=+944.233465756" watchObservedRunningTime="2026-01-06 14:52:27.609218873 +0000 UTC m=+944.236685191" Jan 06 14:52:29 crc kubenswrapper[4744]: I0106 14:52:29.609290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" event={"ID":"c4447713-2058-42f3-9b7f-0225267625e3","Type":"ContainerStarted","Data":"97e28d9e98001f8eff7a02e2fee7d316b8cd46feff636cbeb8d6d3589cc2ba0b"} Jan 06 14:52:29 crc kubenswrapper[4744]: I0106 14:52:29.609746 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:43 crc kubenswrapper[4744]: I0106 14:52:43.212484 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" Jan 06 14:52:43 crc kubenswrapper[4744]: I0106 14:52:43.237885 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-74784fd878-9zcxz" podStartSLOduration=15.874160471 podStartE2EDuration="21.23786389s" podCreationTimestamp="2026-01-06 14:52:22 +0000 UTC" firstStartedPulling="2026-01-06 14:52:23.668515721 +0000 UTC m=+940.295982039" lastFinishedPulling="2026-01-06 14:52:29.03221914 +0000 UTC m=+945.659685458" observedRunningTime="2026-01-06 14:52:29.641602487 +0000 UTC m=+946.269068845" watchObservedRunningTime="2026-01-06 14:52:43.23786389 +0000 UTC m=+959.865330208" Jan 06 14:53:02 crc kubenswrapper[4744]: I0106 14:53:02.814361 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-744888559c-nspzf" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.561503 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-mmvx7"] Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.564932 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.568416 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.568919 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-qzt9t" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.569127 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.569662 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795"] Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.570736 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.572036 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.575579 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795"] Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.653650 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qzjdj"] Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.655204 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.658441 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.658545 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.662458 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-d4s9h" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.673022 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-wl5kh"] Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.674495 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.674652 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.676979 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.687560 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-wl5kh"] Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.688068 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpsj6\" (UniqueName: \"kubernetes.io/projected/4d920867-2ba6-42f6-acc3-266cbfbc7716-kube-api-access-dpsj6\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.688119 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-frr-conf\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.688283 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f49af91f-01db-4e5f-b442-11f20ec7a829-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-7t795\" (UID: \"f49af91f-01db-4e5f-b442-11f20ec7a829\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.688431 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4d920867-2ba6-42f6-acc3-266cbfbc7716-metrics-certs\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.688491 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4d920867-2ba6-42f6-acc3-266cbfbc7716-frr-startup\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.688596 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-reloader\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.688682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-metrics\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.688709 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn2cv\" (UniqueName: \"kubernetes.io/projected/f49af91f-01db-4e5f-b442-11f20ec7a829-kube-api-access-fn2cv\") pod \"frr-k8s-webhook-server-7784b6fcf-7t795\" (UID: \"f49af91f-01db-4e5f-b442-11f20ec7a829\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.688756 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-frr-sockets\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.789770 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpsj6\" (UniqueName: \"kubernetes.io/projected/4d920867-2ba6-42f6-acc3-266cbfbc7716-kube-api-access-dpsj6\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.789819 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-frr-conf\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.789868 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19bdb736-8dea-45d9-b5cf-f9ee3407713a-metrics-certs\") pod \"controller-5bddd4b946-wl5kh\" (UID: \"19bdb736-8dea-45d9-b5cf-f9ee3407713a\") " pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.789890 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/27d52d30-370e-44a9-82b1-eb0d7e197220-metallb-excludel2\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.789911 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmb72\" (UniqueName: \"kubernetes.io/projected/27d52d30-370e-44a9-82b1-eb0d7e197220-kube-api-access-wmb72\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.789931 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19bdb736-8dea-45d9-b5cf-f9ee3407713a-cert\") pod \"controller-5bddd4b946-wl5kh\" (UID: \"19bdb736-8dea-45d9-b5cf-f9ee3407713a\") " pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790018 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-metrics-certs\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790036 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kz29\" (UniqueName: \"kubernetes.io/projected/19bdb736-8dea-45d9-b5cf-f9ee3407713a-kube-api-access-5kz29\") pod \"controller-5bddd4b946-wl5kh\" (UID: \"19bdb736-8dea-45d9-b5cf-f9ee3407713a\") " pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790056 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f49af91f-01db-4e5f-b442-11f20ec7a829-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-7t795\" (UID: \"f49af91f-01db-4e5f-b442-11f20ec7a829\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790078 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-memberlist\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790103 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4d920867-2ba6-42f6-acc3-266cbfbc7716-metrics-certs\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790120 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4d920867-2ba6-42f6-acc3-266cbfbc7716-frr-startup\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790143 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-reloader\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790180 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-metrics\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790195 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn2cv\" (UniqueName: \"kubernetes.io/projected/f49af91f-01db-4e5f-b442-11f20ec7a829-kube-api-access-fn2cv\") pod \"frr-k8s-webhook-server-7784b6fcf-7t795\" (UID: \"f49af91f-01db-4e5f-b442-11f20ec7a829\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790215 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-frr-sockets\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.790611 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-frr-conf\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.791414 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-frr-sockets\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.791987 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/4d920867-2ba6-42f6-acc3-266cbfbc7716-frr-startup\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.792221 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-metrics\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.792590 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/4d920867-2ba6-42f6-acc3-266cbfbc7716-reloader\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.802872 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f49af91f-01db-4e5f-b442-11f20ec7a829-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-7t795\" (UID: \"f49af91f-01db-4e5f-b442-11f20ec7a829\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.813674 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4d920867-2ba6-42f6-acc3-266cbfbc7716-metrics-certs\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.817836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn2cv\" (UniqueName: \"kubernetes.io/projected/f49af91f-01db-4e5f-b442-11f20ec7a829-kube-api-access-fn2cv\") pod \"frr-k8s-webhook-server-7784b6fcf-7t795\" (UID: \"f49af91f-01db-4e5f-b442-11f20ec7a829\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.825842 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpsj6\" (UniqueName: \"kubernetes.io/projected/4d920867-2ba6-42f6-acc3-266cbfbc7716-kube-api-access-dpsj6\") pod \"frr-k8s-mmvx7\" (UID: \"4d920867-2ba6-42f6-acc3-266cbfbc7716\") " pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.888457 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.891586 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19bdb736-8dea-45d9-b5cf-f9ee3407713a-cert\") pod \"controller-5bddd4b946-wl5kh\" (UID: \"19bdb736-8dea-45d9-b5cf-f9ee3407713a\") " pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.891649 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-metrics-certs\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.891690 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kz29\" (UniqueName: \"kubernetes.io/projected/19bdb736-8dea-45d9-b5cf-f9ee3407713a-kube-api-access-5kz29\") pod \"controller-5bddd4b946-wl5kh\" (UID: \"19bdb736-8dea-45d9-b5cf-f9ee3407713a\") " pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.891718 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-memberlist\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.891857 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19bdb736-8dea-45d9-b5cf-f9ee3407713a-metrics-certs\") pod \"controller-5bddd4b946-wl5kh\" (UID: \"19bdb736-8dea-45d9-b5cf-f9ee3407713a\") " pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.891875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/27d52d30-370e-44a9-82b1-eb0d7e197220-metallb-excludel2\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.891895 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmb72\" (UniqueName: \"kubernetes.io/projected/27d52d30-370e-44a9-82b1-eb0d7e197220-kube-api-access-wmb72\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: E0106 14:53:03.891984 4744 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 06 14:53:03 crc kubenswrapper[4744]: E0106 14:53:03.892042 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-memberlist podName:27d52d30-370e-44a9-82b1-eb0d7e197220 nodeName:}" failed. No retries permitted until 2026-01-06 14:53:04.392025539 +0000 UTC m=+981.019491847 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-memberlist") pod "speaker-qzjdj" (UID: "27d52d30-370e-44a9-82b1-eb0d7e197220") : secret "metallb-memberlist" not found Jan 06 14:53:03 crc kubenswrapper[4744]: E0106 14:53:03.892276 4744 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Jan 06 14:53:03 crc kubenswrapper[4744]: E0106 14:53:03.892306 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-metrics-certs podName:27d52d30-370e-44a9-82b1-eb0d7e197220 nodeName:}" failed. No retries permitted until 2026-01-06 14:53:04.392299046 +0000 UTC m=+981.019765364 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-metrics-certs") pod "speaker-qzjdj" (UID: "27d52d30-370e-44a9-82b1-eb0d7e197220") : secret "speaker-certs-secret" not found Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.892956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/27d52d30-370e-44a9-82b1-eb0d7e197220-metallb-excludel2\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.893181 4744 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.895344 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/19bdb736-8dea-45d9-b5cf-f9ee3407713a-metrics-certs\") pod \"controller-5bddd4b946-wl5kh\" (UID: \"19bdb736-8dea-45d9-b5cf-f9ee3407713a\") " pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.895584 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.908300 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/19bdb736-8dea-45d9-b5cf-f9ee3407713a-cert\") pod \"controller-5bddd4b946-wl5kh\" (UID: \"19bdb736-8dea-45d9-b5cf-f9ee3407713a\") " pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.915318 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmb72\" (UniqueName: \"kubernetes.io/projected/27d52d30-370e-44a9-82b1-eb0d7e197220-kube-api-access-wmb72\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.916467 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kz29\" (UniqueName: \"kubernetes.io/projected/19bdb736-8dea-45d9-b5cf-f9ee3407713a-kube-api-access-5kz29\") pod \"controller-5bddd4b946-wl5kh\" (UID: \"19bdb736-8dea-45d9-b5cf-f9ee3407713a\") " pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:03 crc kubenswrapper[4744]: I0106 14:53:03.991535 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.340337 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795"] Jan 06 14:53:04 crc kubenswrapper[4744]: W0106 14:53:04.344101 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf49af91f_01db_4e5f_b442_11f20ec7a829.slice/crio-27a720e155c80d6be995edbffef2fb17e1d95f2575f236f3bad723f18558ce8b WatchSource:0}: Error finding container 27a720e155c80d6be995edbffef2fb17e1d95f2575f236f3bad723f18558ce8b: Status 404 returned error can't find the container with id 27a720e155c80d6be995edbffef2fb17e1d95f2575f236f3bad723f18558ce8b Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.403759 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-metrics-certs\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.403815 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-memberlist\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:04 crc kubenswrapper[4744]: E0106 14:53:04.404283 4744 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 06 14:53:04 crc kubenswrapper[4744]: E0106 14:53:04.404374 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-memberlist podName:27d52d30-370e-44a9-82b1-eb0d7e197220 nodeName:}" failed. No retries permitted until 2026-01-06 14:53:05.404354829 +0000 UTC m=+982.031821137 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-memberlist") pod "speaker-qzjdj" (UID: "27d52d30-370e-44a9-82b1-eb0d7e197220") : secret "metallb-memberlist" not found Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.420322 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-metrics-certs\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.421326 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-wl5kh"] Jan 06 14:53:04 crc kubenswrapper[4744]: W0106 14:53:04.425789 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19bdb736_8dea_45d9_b5cf_f9ee3407713a.slice/crio-dc301db37fdda83d77bb26448379026e58861b3e6042099caa3cff786530e50b WatchSource:0}: Error finding container dc301db37fdda83d77bb26448379026e58861b3e6042099caa3cff786530e50b: Status 404 returned error can't find the container with id dc301db37fdda83d77bb26448379026e58861b3e6042099caa3cff786530e50b Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.932938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-wl5kh" event={"ID":"19bdb736-8dea-45d9-b5cf-f9ee3407713a","Type":"ContainerStarted","Data":"8dc2607148aabe97108bc9230f376f25b0014c3e3f9c38b9fbe4f3a73368934a"} Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.933339 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.933353 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-wl5kh" event={"ID":"19bdb736-8dea-45d9-b5cf-f9ee3407713a","Type":"ContainerStarted","Data":"f8c9aaae8d1ecbf59705c9e618212eba38117e712f6bacb92edfe0a0ec3e616f"} Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.933363 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-wl5kh" event={"ID":"19bdb736-8dea-45d9-b5cf-f9ee3407713a","Type":"ContainerStarted","Data":"dc301db37fdda83d77bb26448379026e58861b3e6042099caa3cff786530e50b"} Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.934746 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerStarted","Data":"c45b9dd95db2f450b742fcefc691fe657c11b5272a784d2537dad8a2d98254eb"} Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.936277 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" event={"ID":"f49af91f-01db-4e5f-b442-11f20ec7a829","Type":"ContainerStarted","Data":"27a720e155c80d6be995edbffef2fb17e1d95f2575f236f3bad723f18558ce8b"} Jan 06 14:53:04 crc kubenswrapper[4744]: I0106 14:53:04.955368 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-wl5kh" podStartSLOduration=1.955339293 podStartE2EDuration="1.955339293s" podCreationTimestamp="2026-01-06 14:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:53:04.948428812 +0000 UTC m=+981.575895190" watchObservedRunningTime="2026-01-06 14:53:04.955339293 +0000 UTC m=+981.582805651" Jan 06 14:53:05 crc kubenswrapper[4744]: I0106 14:53:05.424006 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-memberlist\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:05 crc kubenswrapper[4744]: I0106 14:53:05.430247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/27d52d30-370e-44a9-82b1-eb0d7e197220-memberlist\") pod \"speaker-qzjdj\" (UID: \"27d52d30-370e-44a9-82b1-eb0d7e197220\") " pod="metallb-system/speaker-qzjdj" Jan 06 14:53:05 crc kubenswrapper[4744]: I0106 14:53:05.473050 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qzjdj" Jan 06 14:53:05 crc kubenswrapper[4744]: W0106 14:53:05.531319 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27d52d30_370e_44a9_82b1_eb0d7e197220.slice/crio-1da405ef529d76d4e4b942af18052a1c732a08c348fbd1a8f2d1e72a2e404574 WatchSource:0}: Error finding container 1da405ef529d76d4e4b942af18052a1c732a08c348fbd1a8f2d1e72a2e404574: Status 404 returned error can't find the container with id 1da405ef529d76d4e4b942af18052a1c732a08c348fbd1a8f2d1e72a2e404574 Jan 06 14:53:05 crc kubenswrapper[4744]: I0106 14:53:05.952204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qzjdj" event={"ID":"27d52d30-370e-44a9-82b1-eb0d7e197220","Type":"ContainerStarted","Data":"58f330ade22100fcfcf29f04de3048fc45deb7c2dbca9c63916830eb56e97399"} Jan 06 14:53:05 crc kubenswrapper[4744]: I0106 14:53:05.952251 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qzjdj" event={"ID":"27d52d30-370e-44a9-82b1-eb0d7e197220","Type":"ContainerStarted","Data":"1da405ef529d76d4e4b942af18052a1c732a08c348fbd1a8f2d1e72a2e404574"} Jan 06 14:53:06 crc kubenswrapper[4744]: I0106 14:53:06.975037 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qzjdj" event={"ID":"27d52d30-370e-44a9-82b1-eb0d7e197220","Type":"ContainerStarted","Data":"fb2bd33c5269bd2ce388d3e28748dcaff49d884f7dc32a168970eb80c6dd3569"} Jan 06 14:53:06 crc kubenswrapper[4744]: I0106 14:53:06.975481 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qzjdj" Jan 06 14:53:07 crc kubenswrapper[4744]: I0106 14:53:07.000515 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qzjdj" podStartSLOduration=4.000498693 podStartE2EDuration="4.000498693s" podCreationTimestamp="2026-01-06 14:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:53:06.998380297 +0000 UTC m=+983.625846615" watchObservedRunningTime="2026-01-06 14:53:07.000498693 +0000 UTC m=+983.627965011" Jan 06 14:53:12 crc kubenswrapper[4744]: I0106 14:53:12.021525 4744 generic.go:334] "Generic (PLEG): container finished" podID="4d920867-2ba6-42f6-acc3-266cbfbc7716" containerID="5966383d1af62afb83b4f2425b03228cf395ad84bdfdccd416b689fa35b5ae3c" exitCode=0 Jan 06 14:53:12 crc kubenswrapper[4744]: I0106 14:53:12.021597 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerDied","Data":"5966383d1af62afb83b4f2425b03228cf395ad84bdfdccd416b689fa35b5ae3c"} Jan 06 14:53:12 crc kubenswrapper[4744]: I0106 14:53:12.025606 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" event={"ID":"f49af91f-01db-4e5f-b442-11f20ec7a829","Type":"ContainerStarted","Data":"2ad81ce0e0d8b3f9d553c5dfa2795c0e67373aea6a309419095552fd4cb2c301"} Jan 06 14:53:12 crc kubenswrapper[4744]: I0106 14:53:12.026302 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:12 crc kubenswrapper[4744]: I0106 14:53:12.080335 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" podStartSLOduration=1.792370295 podStartE2EDuration="9.080306489s" podCreationTimestamp="2026-01-06 14:53:03 +0000 UTC" firstStartedPulling="2026-01-06 14:53:04.346298905 +0000 UTC m=+980.973765223" lastFinishedPulling="2026-01-06 14:53:11.634235089 +0000 UTC m=+988.261701417" observedRunningTime="2026-01-06 14:53:12.077931127 +0000 UTC m=+988.705397515" watchObservedRunningTime="2026-01-06 14:53:12.080306489 +0000 UTC m=+988.707772847" Jan 06 14:53:13 crc kubenswrapper[4744]: I0106 14:53:13.037520 4744 generic.go:334] "Generic (PLEG): container finished" podID="4d920867-2ba6-42f6-acc3-266cbfbc7716" containerID="11ab66b58eedc89d15ebf5ced809c3cf5e86aedd4422b9c97d473c7732c4aa02" exitCode=0 Jan 06 14:53:13 crc kubenswrapper[4744]: I0106 14:53:13.037625 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerDied","Data":"11ab66b58eedc89d15ebf5ced809c3cf5e86aedd4422b9c97d473c7732c4aa02"} Jan 06 14:53:14 crc kubenswrapper[4744]: I0106 14:53:14.051016 4744 generic.go:334] "Generic (PLEG): container finished" podID="4d920867-2ba6-42f6-acc3-266cbfbc7716" containerID="971e36be89ef7193b24d1a7137564a2e1fac31fbfc58f8ec8efe9fffc6ec8ed7" exitCode=0 Jan 06 14:53:14 crc kubenswrapper[4744]: I0106 14:53:14.051092 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerDied","Data":"971e36be89ef7193b24d1a7137564a2e1fac31fbfc58f8ec8efe9fffc6ec8ed7"} Jan 06 14:53:15 crc kubenswrapper[4744]: I0106 14:53:15.067818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerStarted","Data":"56a7e8c9e1bc749e789ef2979f71afc5dca78213cb142b3ebcb0f4ea626e5237"} Jan 06 14:53:15 crc kubenswrapper[4744]: I0106 14:53:15.068107 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerStarted","Data":"11dd3c43148ac2bdfb12adf595bd95810fee12ca12c64f81df90b1007e17876f"} Jan 06 14:53:15 crc kubenswrapper[4744]: I0106 14:53:15.068119 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerStarted","Data":"78fb13368c35b769de6ae8b17f5c4e40c10c4bc351c89f5a44ebe2e2928e4cf3"} Jan 06 14:53:15 crc kubenswrapper[4744]: I0106 14:53:15.068127 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerStarted","Data":"bc71b432a284e0b8d56df6731c6d2de40e9021a35498e1f9fa9a324269c38530"} Jan 06 14:53:15 crc kubenswrapper[4744]: I0106 14:53:15.477228 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qzjdj" Jan 06 14:53:16 crc kubenswrapper[4744]: I0106 14:53:16.084571 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerStarted","Data":"b5027006c3463431ef7938db3bd36c4d8e679e5304b642d0ce588639dbcd6f32"} Jan 06 14:53:16 crc kubenswrapper[4744]: I0106 14:53:16.084621 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mmvx7" event={"ID":"4d920867-2ba6-42f6-acc3-266cbfbc7716","Type":"ContainerStarted","Data":"32df571988a6718af0e42ffdb80eca3ca2eb8b607ef0d77f51e9082203c2221c"} Jan 06 14:53:16 crc kubenswrapper[4744]: I0106 14:53:16.084747 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:16 crc kubenswrapper[4744]: I0106 14:53:16.107922 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-mmvx7" podStartSLOduration=5.6035538559999996 podStartE2EDuration="13.107906862s" podCreationTimestamp="2026-01-06 14:53:03 +0000 UTC" firstStartedPulling="2026-01-06 14:53:04.089286287 +0000 UTC m=+980.716752605" lastFinishedPulling="2026-01-06 14:53:11.593639293 +0000 UTC m=+988.221105611" observedRunningTime="2026-01-06 14:53:16.106602818 +0000 UTC m=+992.734069146" watchObservedRunningTime="2026-01-06 14:53:16.107906862 +0000 UTC m=+992.735373180" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.645533 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-ql6b9"] Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.647765 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ql6b9" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.664736 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.671739 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-spgzl" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.671776 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.681634 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ql6b9"] Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.757547 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8jg4\" (UniqueName: \"kubernetes.io/projected/585efc6e-65d4-4b0c-a103-74cb5f664cbd-kube-api-access-s8jg4\") pod \"openstack-operator-index-ql6b9\" (UID: \"585efc6e-65d4-4b0c-a103-74cb5f664cbd\") " pod="openstack-operators/openstack-operator-index-ql6b9" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.859624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8jg4\" (UniqueName: \"kubernetes.io/projected/585efc6e-65d4-4b0c-a103-74cb5f664cbd-kube-api-access-s8jg4\") pod \"openstack-operator-index-ql6b9\" (UID: \"585efc6e-65d4-4b0c-a103-74cb5f664cbd\") " pod="openstack-operators/openstack-operator-index-ql6b9" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.882946 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8jg4\" (UniqueName: \"kubernetes.io/projected/585efc6e-65d4-4b0c-a103-74cb5f664cbd-kube-api-access-s8jg4\") pod \"openstack-operator-index-ql6b9\" (UID: \"585efc6e-65d4-4b0c-a103-74cb5f664cbd\") " pod="openstack-operators/openstack-operator-index-ql6b9" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.889153 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.927310 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:18 crc kubenswrapper[4744]: I0106 14:53:18.981310 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ql6b9" Jan 06 14:53:19 crc kubenswrapper[4744]: I0106 14:53:19.476222 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-ql6b9"] Jan 06 14:53:19 crc kubenswrapper[4744]: W0106 14:53:19.488307 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod585efc6e_65d4_4b0c_a103_74cb5f664cbd.slice/crio-25f856035599e3a82073b7bb926273b851f8287fd94afcd25dcf0f92414c3533 WatchSource:0}: Error finding container 25f856035599e3a82073b7bb926273b851f8287fd94afcd25dcf0f92414c3533: Status 404 returned error can't find the container with id 25f856035599e3a82073b7bb926273b851f8287fd94afcd25dcf0f92414c3533 Jan 06 14:53:20 crc kubenswrapper[4744]: I0106 14:53:20.118248 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ql6b9" event={"ID":"585efc6e-65d4-4b0c-a103-74cb5f664cbd","Type":"ContainerStarted","Data":"25f856035599e3a82073b7bb926273b851f8287fd94afcd25dcf0f92414c3533"} Jan 06 14:53:21 crc kubenswrapper[4744]: I0106 14:53:21.808579 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-ql6b9"] Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.137623 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ql6b9" event={"ID":"585efc6e-65d4-4b0c-a103-74cb5f664cbd","Type":"ContainerStarted","Data":"2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8"} Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.137783 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-ql6b9" podUID="585efc6e-65d4-4b0c-a103-74cb5f664cbd" containerName="registry-server" containerID="cri-o://2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8" gracePeriod=2 Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.158251 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-ql6b9" podStartSLOduration=1.731360177 podStartE2EDuration="4.158231117s" podCreationTimestamp="2026-01-06 14:53:18 +0000 UTC" firstStartedPulling="2026-01-06 14:53:19.490891204 +0000 UTC m=+996.118357522" lastFinishedPulling="2026-01-06 14:53:21.917762124 +0000 UTC m=+998.545228462" observedRunningTime="2026-01-06 14:53:22.153032471 +0000 UTC m=+998.780498799" watchObservedRunningTime="2026-01-06 14:53:22.158231117 +0000 UTC m=+998.785697435" Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.434828 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7hrfh"] Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.436068 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7hrfh" Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.469301 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7hrfh"] Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.557511 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzn8f\" (UniqueName: \"kubernetes.io/projected/458bdba1-55cb-49ad-81ef-5e4c8b07c920-kube-api-access-hzn8f\") pod \"openstack-operator-index-7hrfh\" (UID: \"458bdba1-55cb-49ad-81ef-5e4c8b07c920\") " pod="openstack-operators/openstack-operator-index-7hrfh" Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.626499 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ql6b9" Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.659330 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8jg4\" (UniqueName: \"kubernetes.io/projected/585efc6e-65d4-4b0c-a103-74cb5f664cbd-kube-api-access-s8jg4\") pod \"585efc6e-65d4-4b0c-a103-74cb5f664cbd\" (UID: \"585efc6e-65d4-4b0c-a103-74cb5f664cbd\") " Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.659974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzn8f\" (UniqueName: \"kubernetes.io/projected/458bdba1-55cb-49ad-81ef-5e4c8b07c920-kube-api-access-hzn8f\") pod \"openstack-operator-index-7hrfh\" (UID: \"458bdba1-55cb-49ad-81ef-5e4c8b07c920\") " pod="openstack-operators/openstack-operator-index-7hrfh" Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.665977 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/585efc6e-65d4-4b0c-a103-74cb5f664cbd-kube-api-access-s8jg4" (OuterVolumeSpecName: "kube-api-access-s8jg4") pod "585efc6e-65d4-4b0c-a103-74cb5f664cbd" (UID: "585efc6e-65d4-4b0c-a103-74cb5f664cbd"). InnerVolumeSpecName "kube-api-access-s8jg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.679012 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzn8f\" (UniqueName: \"kubernetes.io/projected/458bdba1-55cb-49ad-81ef-5e4c8b07c920-kube-api-access-hzn8f\") pod \"openstack-operator-index-7hrfh\" (UID: \"458bdba1-55cb-49ad-81ef-5e4c8b07c920\") " pod="openstack-operators/openstack-operator-index-7hrfh" Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.762211 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8jg4\" (UniqueName: \"kubernetes.io/projected/585efc6e-65d4-4b0c-a103-74cb5f664cbd-kube-api-access-s8jg4\") on node \"crc\" DevicePath \"\"" Jan 06 14:53:22 crc kubenswrapper[4744]: I0106 14:53:22.800631 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7hrfh" Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.146722 4744 generic.go:334] "Generic (PLEG): container finished" podID="585efc6e-65d4-4b0c-a103-74cb5f664cbd" containerID="2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8" exitCode=0 Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.146769 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-ql6b9" Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.146789 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ql6b9" event={"ID":"585efc6e-65d4-4b0c-a103-74cb5f664cbd","Type":"ContainerDied","Data":"2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8"} Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.147236 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-ql6b9" event={"ID":"585efc6e-65d4-4b0c-a103-74cb5f664cbd","Type":"ContainerDied","Data":"25f856035599e3a82073b7bb926273b851f8287fd94afcd25dcf0f92414c3533"} Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.147257 4744 scope.go:117] "RemoveContainer" containerID="2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8" Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.184129 4744 scope.go:117] "RemoveContainer" containerID="2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8" Jan 06 14:53:23 crc kubenswrapper[4744]: E0106 14:53:23.184457 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8\": container with ID starting with 2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8 not found: ID does not exist" containerID="2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8" Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.184504 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8"} err="failed to get container status \"2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8\": rpc error: code = NotFound desc = could not find container \"2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8\": container with ID starting with 2d76e414452415ce20d8b1053cf9de14e78e712662d10a3169bf9d02cf1ca0f8 not found: ID does not exist" Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.194334 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-ql6b9"] Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.202011 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-ql6b9"] Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.280564 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7hrfh"] Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.721151 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="585efc6e-65d4-4b0c-a103-74cb5f664cbd" path="/var/lib/kubelet/pods/585efc6e-65d4-4b0c-a103-74cb5f664cbd/volumes" Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.903019 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-7t795" Jan 06 14:53:23 crc kubenswrapper[4744]: I0106 14:53:23.998069 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-wl5kh" Jan 06 14:53:24 crc kubenswrapper[4744]: I0106 14:53:24.157384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7hrfh" event={"ID":"458bdba1-55cb-49ad-81ef-5e4c8b07c920","Type":"ContainerStarted","Data":"2557b1623bae427b28eb003bb8ca5f2c2e906710a7230579d8e6098b6876e7ce"} Jan 06 14:53:24 crc kubenswrapper[4744]: I0106 14:53:24.157443 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7hrfh" event={"ID":"458bdba1-55cb-49ad-81ef-5e4c8b07c920","Type":"ContainerStarted","Data":"a0e9c65ad0d0fcad023b4f5b65e4a8a0aac375bc66f04cd0ebc9df8ec766c4bb"} Jan 06 14:53:24 crc kubenswrapper[4744]: I0106 14:53:24.173986 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7hrfh" podStartSLOduration=2.038621111 podStartE2EDuration="2.173964554s" podCreationTimestamp="2026-01-06 14:53:22 +0000 UTC" firstStartedPulling="2026-01-06 14:53:23.300382891 +0000 UTC m=+999.927849209" lastFinishedPulling="2026-01-06 14:53:23.435726334 +0000 UTC m=+1000.063192652" observedRunningTime="2026-01-06 14:53:24.170016821 +0000 UTC m=+1000.797483139" watchObservedRunningTime="2026-01-06 14:53:24.173964554 +0000 UTC m=+1000.801430872" Jan 06 14:53:32 crc kubenswrapper[4744]: I0106 14:53:32.800816 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-7hrfh" Jan 06 14:53:32 crc kubenswrapper[4744]: I0106 14:53:32.801360 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-7hrfh" Jan 06 14:53:32 crc kubenswrapper[4744]: I0106 14:53:32.840582 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-7hrfh" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.269535 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-7hrfh" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.873129 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq"] Jan 06 14:53:33 crc kubenswrapper[4744]: E0106 14:53:33.873520 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="585efc6e-65d4-4b0c-a103-74cb5f664cbd" containerName="registry-server" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.873535 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="585efc6e-65d4-4b0c-a103-74cb5f664cbd" containerName="registry-server" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.873783 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="585efc6e-65d4-4b0c-a103-74cb5f664cbd" containerName="registry-server" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.875083 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.877667 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-jtl4b" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.892712 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq"] Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.893998 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-mmvx7" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.969180 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-bundle\") pod \"cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.969547 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-util\") pod \"cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:33 crc kubenswrapper[4744]: I0106 14:53:33.969744 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzfph\" (UniqueName: \"kubernetes.io/projected/f1b99747-da20-401e-a710-13ea2d33e7fc-kube-api-access-lzfph\") pod \"cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:34 crc kubenswrapper[4744]: I0106 14:53:34.072117 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-bundle\") pod \"cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:34 crc kubenswrapper[4744]: I0106 14:53:34.072211 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-util\") pod \"cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:34 crc kubenswrapper[4744]: I0106 14:53:34.072313 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzfph\" (UniqueName: \"kubernetes.io/projected/f1b99747-da20-401e-a710-13ea2d33e7fc-kube-api-access-lzfph\") pod \"cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:34 crc kubenswrapper[4744]: I0106 14:53:34.072804 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-bundle\") pod \"cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:34 crc kubenswrapper[4744]: I0106 14:53:34.072838 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-util\") pod \"cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:34 crc kubenswrapper[4744]: I0106 14:53:34.104199 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzfph\" (UniqueName: \"kubernetes.io/projected/f1b99747-da20-401e-a710-13ea2d33e7fc-kube-api-access-lzfph\") pod \"cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:34 crc kubenswrapper[4744]: I0106 14:53:34.194206 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:34 crc kubenswrapper[4744]: I0106 14:53:34.737042 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq"] Jan 06 14:53:35 crc kubenswrapper[4744]: I0106 14:53:35.254203 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" event={"ID":"f1b99747-da20-401e-a710-13ea2d33e7fc","Type":"ContainerStarted","Data":"d50ff202afba319606742fb83cef76ebb159bbd931ceabc3a5e8170a8d2227a1"} Jan 06 14:53:37 crc kubenswrapper[4744]: I0106 14:53:37.285774 4744 generic.go:334] "Generic (PLEG): container finished" podID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerID="e7e3613d0eb9f2b50b382e46cce3605810fec5a86eb3901c893a07f1606c77bb" exitCode=0 Jan 06 14:53:37 crc kubenswrapper[4744]: I0106 14:53:37.285872 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" event={"ID":"f1b99747-da20-401e-a710-13ea2d33e7fc","Type":"ContainerDied","Data":"e7e3613d0eb9f2b50b382e46cce3605810fec5a86eb3901c893a07f1606c77bb"} Jan 06 14:53:38 crc kubenswrapper[4744]: I0106 14:53:38.303678 4744 generic.go:334] "Generic (PLEG): container finished" podID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerID="833c5697af71426664d9f3eb24cbd41503043a3c2c0edaff9659cc978b0893c2" exitCode=0 Jan 06 14:53:38 crc kubenswrapper[4744]: I0106 14:53:38.303822 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" event={"ID":"f1b99747-da20-401e-a710-13ea2d33e7fc","Type":"ContainerDied","Data":"833c5697af71426664d9f3eb24cbd41503043a3c2c0edaff9659cc978b0893c2"} Jan 06 14:53:39 crc kubenswrapper[4744]: I0106 14:53:39.325285 4744 generic.go:334] "Generic (PLEG): container finished" podID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerID="62b3cdb781dab0f6e6fe1850c6f6750f4a134efe1bb12f81e094d2acf4677de6" exitCode=0 Jan 06 14:53:39 crc kubenswrapper[4744]: I0106 14:53:39.325358 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" event={"ID":"f1b99747-da20-401e-a710-13ea2d33e7fc","Type":"ContainerDied","Data":"62b3cdb781dab0f6e6fe1850c6f6750f4a134efe1bb12f81e094d2acf4677de6"} Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.785876 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.801103 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-bundle\") pod \"f1b99747-da20-401e-a710-13ea2d33e7fc\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.801461 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-util\") pod \"f1b99747-da20-401e-a710-13ea2d33e7fc\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.801516 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzfph\" (UniqueName: \"kubernetes.io/projected/f1b99747-da20-401e-a710-13ea2d33e7fc-kube-api-access-lzfph\") pod \"f1b99747-da20-401e-a710-13ea2d33e7fc\" (UID: \"f1b99747-da20-401e-a710-13ea2d33e7fc\") " Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.805732 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-bundle" (OuterVolumeSpecName: "bundle") pod "f1b99747-da20-401e-a710-13ea2d33e7fc" (UID: "f1b99747-da20-401e-a710-13ea2d33e7fc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.812411 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1b99747-da20-401e-a710-13ea2d33e7fc-kube-api-access-lzfph" (OuterVolumeSpecName: "kube-api-access-lzfph") pod "f1b99747-da20-401e-a710-13ea2d33e7fc" (UID: "f1b99747-da20-401e-a710-13ea2d33e7fc"). InnerVolumeSpecName "kube-api-access-lzfph". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.836361 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-util" (OuterVolumeSpecName: "util") pod "f1b99747-da20-401e-a710-13ea2d33e7fc" (UID: "f1b99747-da20-401e-a710-13ea2d33e7fc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.904938 4744 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-util\") on node \"crc\" DevicePath \"\"" Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.904988 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzfph\" (UniqueName: \"kubernetes.io/projected/f1b99747-da20-401e-a710-13ea2d33e7fc-kube-api-access-lzfph\") on node \"crc\" DevicePath \"\"" Jan 06 14:53:40 crc kubenswrapper[4744]: I0106 14:53:40.905200 4744 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f1b99747-da20-401e-a710-13ea2d33e7fc-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:53:41 crc kubenswrapper[4744]: I0106 14:53:41.346412 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" event={"ID":"f1b99747-da20-401e-a710-13ea2d33e7fc","Type":"ContainerDied","Data":"d50ff202afba319606742fb83cef76ebb159bbd931ceabc3a5e8170a8d2227a1"} Jan 06 14:53:41 crc kubenswrapper[4744]: I0106 14:53:41.346696 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d50ff202afba319606742fb83cef76ebb159bbd931ceabc3a5e8170a8d2227a1" Jan 06 14:53:41 crc kubenswrapper[4744]: I0106 14:53:41.346717 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.579261 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht"] Jan 06 14:53:45 crc kubenswrapper[4744]: E0106 14:53:45.579794 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerName="util" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.579806 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerName="util" Jan 06 14:53:45 crc kubenswrapper[4744]: E0106 14:53:45.579833 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerName="pull" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.579839 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerName="pull" Jan 06 14:53:45 crc kubenswrapper[4744]: E0106 14:53:45.579849 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerName="extract" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.579855 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerName="extract" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.579990 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1b99747-da20-401e-a710-13ea2d33e7fc" containerName="extract" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.580714 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.591023 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-xc889" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.613975 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht"] Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.684650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ss9h\" (UniqueName: \"kubernetes.io/projected/7be82a54-010f-4102-bea3-f3ce988e65b6-kube-api-access-6ss9h\") pod \"openstack-operator-controller-operator-749c9566f9-kw8ht\" (UID: \"7be82a54-010f-4102-bea3-f3ce988e65b6\") " pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.786794 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ss9h\" (UniqueName: \"kubernetes.io/projected/7be82a54-010f-4102-bea3-f3ce988e65b6-kube-api-access-6ss9h\") pod \"openstack-operator-controller-operator-749c9566f9-kw8ht\" (UID: \"7be82a54-010f-4102-bea3-f3ce988e65b6\") " pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.819513 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ss9h\" (UniqueName: \"kubernetes.io/projected/7be82a54-010f-4102-bea3-f3ce988e65b6-kube-api-access-6ss9h\") pod \"openstack-operator-controller-operator-749c9566f9-kw8ht\" (UID: \"7be82a54-010f-4102-bea3-f3ce988e65b6\") " pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" Jan 06 14:53:45 crc kubenswrapper[4744]: I0106 14:53:45.899576 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" Jan 06 14:53:46 crc kubenswrapper[4744]: I0106 14:53:46.380349 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht"] Jan 06 14:53:46 crc kubenswrapper[4744]: W0106 14:53:46.388150 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7be82a54_010f_4102_bea3_f3ce988e65b6.slice/crio-623e6b0e196b4e81800af433d43c92b7854b36084a69e0dc0a31b762ff80822c WatchSource:0}: Error finding container 623e6b0e196b4e81800af433d43c92b7854b36084a69e0dc0a31b762ff80822c: Status 404 returned error can't find the container with id 623e6b0e196b4e81800af433d43c92b7854b36084a69e0dc0a31b762ff80822c Jan 06 14:53:47 crc kubenswrapper[4744]: I0106 14:53:47.401503 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" event={"ID":"7be82a54-010f-4102-bea3-f3ce988e65b6","Type":"ContainerStarted","Data":"623e6b0e196b4e81800af433d43c92b7854b36084a69e0dc0a31b762ff80822c"} Jan 06 14:53:50 crc kubenswrapper[4744]: I0106 14:53:50.427215 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" event={"ID":"7be82a54-010f-4102-bea3-f3ce988e65b6","Type":"ContainerStarted","Data":"dbfa7527cb7451716bf4472f8f18471ffdd717c5ff40734da046589abdc51954"} Jan 06 14:53:50 crc kubenswrapper[4744]: I0106 14:53:50.427832 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" Jan 06 14:53:50 crc kubenswrapper[4744]: I0106 14:53:50.467530 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" podStartSLOduration=1.627079891 podStartE2EDuration="5.46750552s" podCreationTimestamp="2026-01-06 14:53:45 +0000 UTC" firstStartedPulling="2026-01-06 14:53:46.390373537 +0000 UTC m=+1023.017839855" lastFinishedPulling="2026-01-06 14:53:50.230799166 +0000 UTC m=+1026.858265484" observedRunningTime="2026-01-06 14:53:50.460938268 +0000 UTC m=+1027.088404616" watchObservedRunningTime="2026-01-06 14:53:50.46750552 +0000 UTC m=+1027.094971858" Jan 06 14:53:55 crc kubenswrapper[4744]: I0106 14:53:55.902389 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-749c9566f9-kw8ht" Jan 06 14:54:14 crc kubenswrapper[4744]: I0106 14:54:14.423660 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:54:14 crc kubenswrapper[4744]: I0106 14:54:14.424355 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:54:14 crc kubenswrapper[4744]: I0106 14:54:14.986766 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk"] Jan 06 14:54:14 crc kubenswrapper[4744]: I0106 14:54:14.988136 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" Jan 06 14:54:14 crc kubenswrapper[4744]: I0106 14:54:14.989956 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-vbnj5" Jan 06 14:54:14 crc kubenswrapper[4744]: I0106 14:54:14.999626 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.000882 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.002475 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-kdr92" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.005695 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.014247 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.020459 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.021362 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.023009 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-9cf2c" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.038523 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.046231 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.047553 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.053686 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-tvhb2" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.080144 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.081489 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.088571 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-46jnq" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.089211 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.126211 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.142225 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.143298 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.159537 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-8pv5f" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.161015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t76mb\" (UniqueName: \"kubernetes.io/projected/5b589b5b-5f62-4e3d-808b-8c404d1432ac-kube-api-access-t76mb\") pod \"cinder-operator-controller-manager-78979fc445-bm4m2\" (UID: \"5b589b5b-5f62-4e3d-808b-8c404d1432ac\") " pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.161053 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42lqs\" (UniqueName: \"kubernetes.io/projected/8421b8a1-9929-43ab-b203-9899c52d2d41-kube-api-access-42lqs\") pod \"heat-operator-controller-manager-658dd65b86-pnxwz\" (UID: \"8421b8a1-9929-43ab-b203-9899c52d2d41\") " pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.161073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlgss\" (UniqueName: \"kubernetes.io/projected/3653611e-a619-401f-a964-70d245bb4957-kube-api-access-mlgss\") pod \"designate-operator-controller-manager-66f8b87655-j8rsd\" (UID: \"3653611e-a619-401f-a964-70d245bb4957\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.161110 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vph9\" (UniqueName: \"kubernetes.io/projected/130c27ae-5c64-4c06-b5e8-9a1aacea630e-kube-api-access-4vph9\") pod \"barbican-operator-controller-manager-f6f74d6db-lhppk\" (UID: \"130c27ae-5c64-4c06-b5e8-9a1aacea630e\") " pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.161188 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-947tn\" (UniqueName: \"kubernetes.io/projected/8035e88d-2a38-4e95-b204-0dc00fc57bfc-kube-api-access-947tn\") pod \"glance-operator-controller-manager-7b549fc966-clgkb\" (UID: \"8035e88d-2a38-4e95-b204-0dc00fc57bfc\") " pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.161541 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.187611 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.188685 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.199584 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.200264 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-bxskh" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.200495 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-568985c78-dhg75"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.201852 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.206568 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4l5lz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.220317 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-568985c78-dhg75"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.252318 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.264029 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-947tn\" (UniqueName: \"kubernetes.io/projected/8035e88d-2a38-4e95-b204-0dc00fc57bfc-kube-api-access-947tn\") pod \"glance-operator-controller-manager-7b549fc966-clgkb\" (UID: \"8035e88d-2a38-4e95-b204-0dc00fc57bfc\") " pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.264113 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gsn7\" (UniqueName: \"kubernetes.io/projected/f1558cc4-7188-42b8-858a-e83606895217-kube-api-access-9gsn7\") pod \"horizon-operator-controller-manager-7f5ddd8d7b-q9k7b\" (UID: \"f1558cc4-7188-42b8-858a-e83606895217\") " pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.264140 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t76mb\" (UniqueName: \"kubernetes.io/projected/5b589b5b-5f62-4e3d-808b-8c404d1432ac-kube-api-access-t76mb\") pod \"cinder-operator-controller-manager-78979fc445-bm4m2\" (UID: \"5b589b5b-5f62-4e3d-808b-8c404d1432ac\") " pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.264180 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42lqs\" (UniqueName: \"kubernetes.io/projected/8421b8a1-9929-43ab-b203-9899c52d2d41-kube-api-access-42lqs\") pod \"heat-operator-controller-manager-658dd65b86-pnxwz\" (UID: \"8421b8a1-9929-43ab-b203-9899c52d2d41\") " pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.264199 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlgss\" (UniqueName: \"kubernetes.io/projected/3653611e-a619-401f-a964-70d245bb4957-kube-api-access-mlgss\") pod \"designate-operator-controller-manager-66f8b87655-j8rsd\" (UID: \"3653611e-a619-401f-a964-70d245bb4957\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.264223 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgx58\" (UniqueName: \"kubernetes.io/projected/749a8719-1599-4225-a5f1-165fbe644aca-kube-api-access-zgx58\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.264243 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.264271 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vph9\" (UniqueName: \"kubernetes.io/projected/130c27ae-5c64-4c06-b5e8-9a1aacea630e-kube-api-access-4vph9\") pod \"barbican-operator-controller-manager-f6f74d6db-lhppk\" (UID: \"130c27ae-5c64-4c06-b5e8-9a1aacea630e\") " pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.290000 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.290861 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.291374 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.291492 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.295608 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-xv2jz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.296377 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-947tn\" (UniqueName: \"kubernetes.io/projected/8035e88d-2a38-4e95-b204-0dc00fc57bfc-kube-api-access-947tn\") pod \"glance-operator-controller-manager-7b549fc966-clgkb\" (UID: \"8035e88d-2a38-4e95-b204-0dc00fc57bfc\") " pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.297354 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.310212 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42lqs\" (UniqueName: \"kubernetes.io/projected/8421b8a1-9929-43ab-b203-9899c52d2d41-kube-api-access-42lqs\") pod \"heat-operator-controller-manager-658dd65b86-pnxwz\" (UID: \"8421b8a1-9929-43ab-b203-9899c52d2d41\") " pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.297144 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-szgnz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.319984 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t76mb\" (UniqueName: \"kubernetes.io/projected/5b589b5b-5f62-4e3d-808b-8c404d1432ac-kube-api-access-t76mb\") pod \"cinder-operator-controller-manager-78979fc445-bm4m2\" (UID: \"5b589b5b-5f62-4e3d-808b-8c404d1432ac\") " pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.323087 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.330751 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vph9\" (UniqueName: \"kubernetes.io/projected/130c27ae-5c64-4c06-b5e8-9a1aacea630e-kube-api-access-4vph9\") pod \"barbican-operator-controller-manager-f6f74d6db-lhppk\" (UID: \"130c27ae-5c64-4c06-b5e8-9a1aacea630e\") " pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.344412 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlgss\" (UniqueName: \"kubernetes.io/projected/3653611e-a619-401f-a964-70d245bb4957-kube-api-access-mlgss\") pod \"designate-operator-controller-manager-66f8b87655-j8rsd\" (UID: \"3653611e-a619-401f-a964-70d245bb4957\") " pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.344475 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.350894 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.366284 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.392685 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.392784 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.394078 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:15 crc kubenswrapper[4744]: E0106 14:54:15.394389 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:15 crc kubenswrapper[4744]: E0106 14:54:15.394453 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert podName:749a8719-1599-4225-a5f1-165fbe644aca nodeName:}" failed. No retries permitted until 2026-01-06 14:54:15.894432179 +0000 UTC m=+1052.521898507 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert") pod "infra-operator-controller-manager-6d99759cf-pgssl" (UID: "749a8719-1599-4225-a5f1-165fbe644aca") : secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.394780 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdkgx\" (UniqueName: \"kubernetes.io/projected/aabafe79-ea6f-4f01-9efb-e1c1a9e2334e-kube-api-access-rdkgx\") pod \"ironic-operator-controller-manager-f99f54bc8-hwxtd\" (UID: \"aabafe79-ea6f-4f01-9efb-e1c1a9e2334e\") " pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.394892 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkllg\" (UniqueName: \"kubernetes.io/projected/66470ebb-8a02-4745-941b-5b6a1b97df1c-kube-api-access-nkllg\") pod \"keystone-operator-controller-manager-568985c78-dhg75\" (UID: \"66470ebb-8a02-4745-941b-5b6a1b97df1c\") " pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.395068 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6ftb\" (UniqueName: \"kubernetes.io/projected/8cc57759-7dc1-4360-8d81-4744355ec4df-kube-api-access-b6ftb\") pod \"manila-operator-controller-manager-598945d5b8-qc2xd\" (UID: \"8cc57759-7dc1-4360-8d81-4744355ec4df\") " pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.395183 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gsn7\" (UniqueName: \"kubernetes.io/projected/f1558cc4-7188-42b8-858a-e83606895217-kube-api-access-9gsn7\") pod \"horizon-operator-controller-manager-7f5ddd8d7b-q9k7b\" (UID: \"f1558cc4-7188-42b8-858a-e83606895217\") " pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.395299 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgx58\" (UniqueName: \"kubernetes.io/projected/749a8719-1599-4225-a5f1-165fbe644aca-kube-api-access-zgx58\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.395628 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-264x6" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.429730 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.483881 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gsn7\" (UniqueName: \"kubernetes.io/projected/f1558cc4-7188-42b8-858a-e83606895217-kube-api-access-9gsn7\") pod \"horizon-operator-controller-manager-7f5ddd8d7b-q9k7b\" (UID: \"f1558cc4-7188-42b8-858a-e83606895217\") " pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.498621 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncsk5\" (UniqueName: \"kubernetes.io/projected/0abaa053-82ac-4b9f-a4d2-2bc5024a97eb-kube-api-access-ncsk5\") pod \"mariadb-operator-controller-manager-7b88bfc995-k8n6t\" (UID: \"0abaa053-82ac-4b9f-a4d2-2bc5024a97eb\") " pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.498687 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdkgx\" (UniqueName: \"kubernetes.io/projected/aabafe79-ea6f-4f01-9efb-e1c1a9e2334e-kube-api-access-rdkgx\") pod \"ironic-operator-controller-manager-f99f54bc8-hwxtd\" (UID: \"aabafe79-ea6f-4f01-9efb-e1c1a9e2334e\") " pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.498718 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkllg\" (UniqueName: \"kubernetes.io/projected/66470ebb-8a02-4745-941b-5b6a1b97df1c-kube-api-access-nkllg\") pod \"keystone-operator-controller-manager-568985c78-dhg75\" (UID: \"66470ebb-8a02-4745-941b-5b6a1b97df1c\") " pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.498782 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6ftb\" (UniqueName: \"kubernetes.io/projected/8cc57759-7dc1-4360-8d81-4744355ec4df-kube-api-access-b6ftb\") pod \"manila-operator-controller-manager-598945d5b8-qc2xd\" (UID: \"8cc57759-7dc1-4360-8d81-4744355ec4df\") " pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.518876 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgx58\" (UniqueName: \"kubernetes.io/projected/749a8719-1599-4225-a5f1-165fbe644aca-kube-api-access-zgx58\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.582878 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkllg\" (UniqueName: \"kubernetes.io/projected/66470ebb-8a02-4745-941b-5b6a1b97df1c-kube-api-access-nkllg\") pod \"keystone-operator-controller-manager-568985c78-dhg75\" (UID: \"66470ebb-8a02-4745-941b-5b6a1b97df1c\") " pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.583623 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdkgx\" (UniqueName: \"kubernetes.io/projected/aabafe79-ea6f-4f01-9efb-e1c1a9e2334e-kube-api-access-rdkgx\") pod \"ironic-operator-controller-manager-f99f54bc8-hwxtd\" (UID: \"aabafe79-ea6f-4f01-9efb-e1c1a9e2334e\") " pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.583692 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.584773 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.586269 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.587903 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-bgpm2" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.589403 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6ftb\" (UniqueName: \"kubernetes.io/projected/8cc57759-7dc1-4360-8d81-4744355ec4df-kube-api-access-b6ftb\") pod \"manila-operator-controller-manager-598945d5b8-qc2xd\" (UID: \"8cc57759-7dc1-4360-8d81-4744355ec4df\") " pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.597367 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.601828 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9m8h\" (UniqueName: \"kubernetes.io/projected/5a327dca-6d91-4a3d-a824-b9f9080338dc-kube-api-access-b9m8h\") pod \"neutron-operator-controller-manager-7cd87b778f-bl6hr\" (UID: \"5a327dca-6d91-4a3d-a824-b9f9080338dc\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.602705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncsk5\" (UniqueName: \"kubernetes.io/projected/0abaa053-82ac-4b9f-a4d2-2bc5024a97eb-kube-api-access-ncsk5\") pod \"mariadb-operator-controller-manager-7b88bfc995-k8n6t\" (UID: \"0abaa053-82ac-4b9f-a4d2-2bc5024a97eb\") " pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.612833 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.619038 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.621512 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.639881 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncsk5\" (UniqueName: \"kubernetes.io/projected/0abaa053-82ac-4b9f-a4d2-2bc5024a97eb-kube-api-access-ncsk5\") pod \"mariadb-operator-controller-manager-7b88bfc995-k8n6t\" (UID: \"0abaa053-82ac-4b9f-a4d2-2bc5024a97eb\") " pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.651271 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.670664 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.681538 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.700974 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.701100 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.703327 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqlbs\" (UniqueName: \"kubernetes.io/projected/70440288-e0e0-42fd-b310-da1267956b3d-kube-api-access-mqlbs\") pod \"nova-operator-controller-manager-5fbbf8b6cc-twhxx\" (UID: \"70440288-e0e0-42fd-b310-da1267956b3d\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.703381 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9m8h\" (UniqueName: \"kubernetes.io/projected/5a327dca-6d91-4a3d-a824-b9f9080338dc-kube-api-access-b9m8h\") pod \"neutron-operator-controller-manager-7cd87b778f-bl6hr\" (UID: \"5a327dca-6d91-4a3d-a824-b9f9080338dc\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.706305 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-4h5g6" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.732738 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9m8h\" (UniqueName: \"kubernetes.io/projected/5a327dca-6d91-4a3d-a824-b9f9080338dc-kube-api-access-b9m8h\") pod \"neutron-operator-controller-manager-7cd87b778f-bl6hr\" (UID: \"5a327dca-6d91-4a3d-a824-b9f9080338dc\") " pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.753577 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.755337 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.755970 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.756569 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.756638 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.757047 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.757426 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.760219 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-2dwjc" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.760434 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-lvf7f" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.760692 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-n88rp" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.769565 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.772995 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.779107 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.783793 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.807277 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.807413 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-ldrmw" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.807473 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.808492 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqlbs\" (UniqueName: \"kubernetes.io/projected/70440288-e0e0-42fd-b310-da1267956b3d-kube-api-access-mqlbs\") pod \"nova-operator-controller-manager-5fbbf8b6cc-twhxx\" (UID: \"70440288-e0e0-42fd-b310-da1267956b3d\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.816768 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.816812 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.816887 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.819329 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-rx6s5" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.826964 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.838866 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.840249 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.843675 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-sr52f" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.849551 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.862993 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.864005 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqlbs\" (UniqueName: \"kubernetes.io/projected/70440288-e0e0-42fd-b310-da1267956b3d-kube-api-access-mqlbs\") pod \"nova-operator-controller-manager-5fbbf8b6cc-twhxx\" (UID: \"70440288-e0e0-42fd-b310-da1267956b3d\") " pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.864074 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.865948 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-hc9nw" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.884066 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.912965 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9f47\" (UniqueName: \"kubernetes.io/projected/73e35a67-364a-487c-a7bf-5e537cbbecbe-kube-api-access-x9f47\") pod \"swift-operator-controller-manager-bb586bbf4-p5kz7\" (UID: \"73e35a67-364a-487c-a7bf-5e537cbbecbe\") " pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.913007 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjv6x\" (UniqueName: \"kubernetes.io/projected/9ef85238-0ee3-44af-98ad-734994758fec-kube-api-access-cjv6x\") pod \"placement-operator-controller-manager-9b6f8f78c-fvltb\" (UID: \"9ef85238-0ee3-44af-98ad-734994758fec\") " pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.913037 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znbg8\" (UniqueName: \"kubernetes.io/projected/65ec28e9-38c5-4972-b8b8-1447de87ee85-kube-api-access-znbg8\") pod \"test-operator-controller-manager-6c866cfdcb-8q8ks\" (UID: \"65ec28e9-38c5-4972-b8b8-1447de87ee85\") " pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.913058 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p447\" (UniqueName: \"kubernetes.io/projected/88292fa9-b45e-4681-97df-e05a423895cc-kube-api-access-8p447\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.913213 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.913260 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shhzp\" (UniqueName: \"kubernetes.io/projected/d18584e4-6488-4fc3-9992-9ded578fd05e-kube-api-access-shhzp\") pod \"telemetry-operator-controller-manager-5649998c8c-g454h\" (UID: \"d18584e4-6488-4fc3-9992-9ded578fd05e\") " pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.913283 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcvtc\" (UniqueName: \"kubernetes.io/projected/96fdaf1c-2733-4272-abda-fb28fa6f6a6a-kube-api-access-mcvtc\") pod \"octavia-operator-controller-manager-68c649d9d-dzwnd\" (UID: \"96fdaf1c-2733-4272-abda-fb28fa6f6a6a\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.913389 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d5ds\" (UniqueName: \"kubernetes.io/projected/36725987-d3b2-4216-bfe0-69fef2d8bfd7-kube-api-access-4d5ds\") pod \"ovn-operator-controller-manager-bf6d4f946-8nnk9\" (UID: \"36725987-d3b2-4216-bfe0-69fef2d8bfd7\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.913419 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:15 crc kubenswrapper[4744]: E0106 14:54:15.917416 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:15 crc kubenswrapper[4744]: E0106 14:54:15.917471 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert podName:749a8719-1599-4225-a5f1-165fbe644aca nodeName:}" failed. No retries permitted until 2026-01-06 14:54:16.91745251 +0000 UTC m=+1053.544918828 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert") pod "infra-operator-controller-manager-6d99759cf-pgssl" (UID: "749a8719-1599-4225-a5f1-165fbe644aca") : secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.923147 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx"] Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.927081 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.930080 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-jz7q6" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.975088 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" Jan 06 14:54:15 crc kubenswrapper[4744]: I0106 14:54:15.984709 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx"] Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.016960 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.017040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shhzp\" (UniqueName: \"kubernetes.io/projected/d18584e4-6488-4fc3-9992-9ded578fd05e-kube-api-access-shhzp\") pod \"telemetry-operator-controller-manager-5649998c8c-g454h\" (UID: \"d18584e4-6488-4fc3-9992-9ded578fd05e\") " pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.017081 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcvtc\" (UniqueName: \"kubernetes.io/projected/96fdaf1c-2733-4272-abda-fb28fa6f6a6a-kube-api-access-mcvtc\") pod \"octavia-operator-controller-manager-68c649d9d-dzwnd\" (UID: \"96fdaf1c-2733-4272-abda-fb28fa6f6a6a\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.017184 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d5ds\" (UniqueName: \"kubernetes.io/projected/36725987-d3b2-4216-bfe0-69fef2d8bfd7-kube-api-access-4d5ds\") pod \"ovn-operator-controller-manager-bf6d4f946-8nnk9\" (UID: \"36725987-d3b2-4216-bfe0-69fef2d8bfd7\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.017240 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9f47\" (UniqueName: \"kubernetes.io/projected/73e35a67-364a-487c-a7bf-5e537cbbecbe-kube-api-access-x9f47\") pod \"swift-operator-controller-manager-bb586bbf4-p5kz7\" (UID: \"73e35a67-364a-487c-a7bf-5e537cbbecbe\") " pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.017274 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjv6x\" (UniqueName: \"kubernetes.io/projected/9ef85238-0ee3-44af-98ad-734994758fec-kube-api-access-cjv6x\") pod \"placement-operator-controller-manager-9b6f8f78c-fvltb\" (UID: \"9ef85238-0ee3-44af-98ad-734994758fec\") " pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.017309 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znbg8\" (UniqueName: \"kubernetes.io/projected/65ec28e9-38c5-4972-b8b8-1447de87ee85-kube-api-access-znbg8\") pod \"test-operator-controller-manager-6c866cfdcb-8q8ks\" (UID: \"65ec28e9-38c5-4972-b8b8-1447de87ee85\") " pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.017337 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p447\" (UniqueName: \"kubernetes.io/projected/88292fa9-b45e-4681-97df-e05a423895cc-kube-api-access-8p447\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.017793 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.017859 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert podName:88292fa9-b45e-4681-97df-e05a423895cc nodeName:}" failed. No retries permitted until 2026-01-06 14:54:16.517841045 +0000 UTC m=+1053.145307363 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert") pod "openstack-baremetal-operator-controller-manager-78948ddfd785mfz" (UID: "88292fa9-b45e-4681-97df-e05a423895cc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.026392 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t"] Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.029262 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.032193 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.033023 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.033363 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.033525 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-fff9l" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.049263 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcvtc\" (UniqueName: \"kubernetes.io/projected/96fdaf1c-2733-4272-abda-fb28fa6f6a6a-kube-api-access-mcvtc\") pod \"octavia-operator-controller-manager-68c649d9d-dzwnd\" (UID: \"96fdaf1c-2733-4272-abda-fb28fa6f6a6a\") " pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.058208 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p447\" (UniqueName: \"kubernetes.io/projected/88292fa9-b45e-4681-97df-e05a423895cc-kube-api-access-8p447\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.058372 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shhzp\" (UniqueName: \"kubernetes.io/projected/d18584e4-6488-4fc3-9992-9ded578fd05e-kube-api-access-shhzp\") pod \"telemetry-operator-controller-manager-5649998c8c-g454h\" (UID: \"d18584e4-6488-4fc3-9992-9ded578fd05e\") " pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.060407 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znbg8\" (UniqueName: \"kubernetes.io/projected/65ec28e9-38c5-4972-b8b8-1447de87ee85-kube-api-access-znbg8\") pod \"test-operator-controller-manager-6c866cfdcb-8q8ks\" (UID: \"65ec28e9-38c5-4972-b8b8-1447de87ee85\") " pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.065794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjv6x\" (UniqueName: \"kubernetes.io/projected/9ef85238-0ee3-44af-98ad-734994758fec-kube-api-access-cjv6x\") pod \"placement-operator-controller-manager-9b6f8f78c-fvltb\" (UID: \"9ef85238-0ee3-44af-98ad-734994758fec\") " pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.082759 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.109247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9f47\" (UniqueName: \"kubernetes.io/projected/73e35a67-364a-487c-a7bf-5e537cbbecbe-kube-api-access-x9f47\") pod \"swift-operator-controller-manager-bb586bbf4-p5kz7\" (UID: \"73e35a67-364a-487c-a7bf-5e537cbbecbe\") " pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.109763 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d5ds\" (UniqueName: \"kubernetes.io/projected/36725987-d3b2-4216-bfe0-69fef2d8bfd7-kube-api-access-4d5ds\") pod \"ovn-operator-controller-manager-bf6d4f946-8nnk9\" (UID: \"36725987-d3b2-4216-bfe0-69fef2d8bfd7\") " pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.119884 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4f8s\" (UniqueName: \"kubernetes.io/projected/7f9fe9dd-bede-4261-9a8a-d23812cffca2-kube-api-access-z4f8s\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.119966 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.120032 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4jkc\" (UniqueName: \"kubernetes.io/projected/1ea5e62f-608a-4fa7-8887-691c1c314a6e-kube-api-access-b4jkc\") pod \"watcher-operator-controller-manager-9dbdf6486-9j2kx\" (UID: \"1ea5e62f-608a-4fa7-8887-691c1c314a6e\") " pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.120106 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.120827 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t"] Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.144541 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.174989 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.203640 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr"] Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.205443 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.213441 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-szv6g" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.214727 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr"] Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.217630 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.221777 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.221899 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4f8s\" (UniqueName: \"kubernetes.io/projected/7f9fe9dd-bede-4261-9a8a-d23812cffca2-kube-api-access-z4f8s\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.221979 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.222047 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4jkc\" (UniqueName: \"kubernetes.io/projected/1ea5e62f-608a-4fa7-8887-691c1c314a6e-kube-api-access-b4jkc\") pod \"watcher-operator-controller-manager-9dbdf6486-9j2kx\" (UID: \"1ea5e62f-608a-4fa7-8887-691c1c314a6e\") " pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.222063 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.222080 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j72f8\" (UniqueName: \"kubernetes.io/projected/fec21671-aa20-4019-bc89-9e0fc135e394-kube-api-access-j72f8\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bjtfr\" (UID: \"fec21671-aa20-4019-bc89-9e0fc135e394\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.222129 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:16.722109677 +0000 UTC m=+1053.349575995 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.222382 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.222406 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:16.722399435 +0000 UTC m=+1053.349865743 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "metrics-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.237790 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.248217 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4f8s\" (UniqueName: \"kubernetes.io/projected/7f9fe9dd-bede-4261-9a8a-d23812cffca2-kube-api-access-z4f8s\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.249030 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4jkc\" (UniqueName: \"kubernetes.io/projected/1ea5e62f-608a-4fa7-8887-691c1c314a6e-kube-api-access-b4jkc\") pod \"watcher-operator-controller-manager-9dbdf6486-9j2kx\" (UID: \"1ea5e62f-608a-4fa7-8887-691c1c314a6e\") " pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.259668 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.275494 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.280682 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2"] Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.293103 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd"] Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.324951 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j72f8\" (UniqueName: \"kubernetes.io/projected/fec21671-aa20-4019-bc89-9e0fc135e394-kube-api-access-j72f8\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bjtfr\" (UID: \"fec21671-aa20-4019-bc89-9e0fc135e394\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.343299 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j72f8\" (UniqueName: \"kubernetes.io/projected/fec21671-aa20-4019-bc89-9e0fc135e394-kube-api-access-j72f8\") pod \"rabbitmq-cluster-operator-manager-668c99d594-bjtfr\" (UID: \"fec21671-aa20-4019-bc89-9e0fc135e394\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.527422 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.527630 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.527679 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert podName:88292fa9-b45e-4681-97df-e05a423895cc nodeName:}" failed. No retries permitted until 2026-01-06 14:54:17.527664907 +0000 UTC m=+1054.155131226 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert") pod "openstack-baremetal-operator-controller-manager-78948ddfd785mfz" (UID: "88292fa9-b45e-4681-97df-e05a423895cc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.533615 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz"] Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.544900 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" Jan 06 14:54:16 crc kubenswrapper[4744]: W0106 14:54:16.550458 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8421b8a1_9929_43ab_b203_9899c52d2d41.slice/crio-5e1b2946b9aacd2b7e1c88b01c37f52e38cec157defebc4a27bec55ed29f6be2 WatchSource:0}: Error finding container 5e1b2946b9aacd2b7e1c88b01c37f52e38cec157defebc4a27bec55ed29f6be2: Status 404 returned error can't find the container with id 5e1b2946b9aacd2b7e1c88b01c37f52e38cec157defebc4a27bec55ed29f6be2 Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.551043 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-568985c78-dhg75"] Jan 06 14:54:16 crc kubenswrapper[4744]: W0106 14:54:16.555974 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66470ebb_8a02_4745_941b_5b6a1b97df1c.slice/crio-c54f342544f98a8cd4f3998e371491dbd0bbe5fcdff7bc1b80bb4cf20c2da8e0 WatchSource:0}: Error finding container c54f342544f98a8cd4f3998e371491dbd0bbe5fcdff7bc1b80bb4cf20c2da8e0: Status 404 returned error can't find the container with id c54f342544f98a8cd4f3998e371491dbd0bbe5fcdff7bc1b80bb4cf20c2da8e0 Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.566467 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb"] Jan 06 14:54:16 crc kubenswrapper[4744]: W0106 14:54:16.567714 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8035e88d_2a38_4e95_b204_0dc00fc57bfc.slice/crio-8043834fc840d83b9ca9aa1a134d971d9765cdf11c953affc5f112e5af12da92 WatchSource:0}: Error finding container 8043834fc840d83b9ca9aa1a134d971d9765cdf11c953affc5f112e5af12da92: Status 404 returned error can't find the container with id 8043834fc840d83b9ca9aa1a134d971d9765cdf11c953affc5f112e5af12da92 Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.733567 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.733987 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.734882 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.734936 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:17.734919278 +0000 UTC m=+1054.362385596 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.735047 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.735114 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:17.735095363 +0000 UTC m=+1054.362561681 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "metrics-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.936277 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.936501 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: E0106 14:54:16.936554 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert podName:749a8719-1599-4225-a5f1-165fbe644aca nodeName:}" failed. No retries permitted until 2026-01-06 14:54:18.936538739 +0000 UTC m=+1055.564005057 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert") pod "infra-operator-controller-manager-6d99759cf-pgssl" (UID: "749a8719-1599-4225-a5f1-165fbe644aca") : secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.992938 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" event={"ID":"8421b8a1-9929-43ab-b203-9899c52d2d41","Type":"ContainerStarted","Data":"5e1b2946b9aacd2b7e1c88b01c37f52e38cec157defebc4a27bec55ed29f6be2"} Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.994842 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" event={"ID":"8035e88d-2a38-4e95-b204-0dc00fc57bfc","Type":"ContainerStarted","Data":"8043834fc840d83b9ca9aa1a134d971d9765cdf11c953affc5f112e5af12da92"} Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.995937 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" event={"ID":"5b589b5b-5f62-4e3d-808b-8c404d1432ac","Type":"ContainerStarted","Data":"1e61d542c415b3ef28630de1f919e326c8513531c846adb14ad949e2b9406a90"} Jan 06 14:54:16 crc kubenswrapper[4744]: I0106 14:54:16.996979 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" event={"ID":"66470ebb-8a02-4745-941b-5b6a1b97df1c","Type":"ContainerStarted","Data":"c54f342544f98a8cd4f3998e371491dbd0bbe5fcdff7bc1b80bb4cf20c2da8e0"} Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.000117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" event={"ID":"3653611e-a619-401f-a964-70d245bb4957","Type":"ContainerStarted","Data":"4bcb535a8d45d984a17c26cd08399ec5c9510a539652f031ec3f1343a332dd1f"} Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.088604 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr"] Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.104108 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b"] Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.111156 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0abaa053_82ac_4b9f_a4d2_2bc5024a97eb.slice/crio-b6b50d7d3ceab7dc6456d9a6fa9030cae50638cba0fde0b4382ffc791f30a961 WatchSource:0}: Error finding container b6b50d7d3ceab7dc6456d9a6fa9030cae50638cba0fde0b4382ffc791f30a961: Status 404 returned error can't find the container with id b6b50d7d3ceab7dc6456d9a6fa9030cae50638cba0fde0b4382ffc791f30a961 Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.112815 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk"] Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.114279 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a327dca_6d91_4a3d_a824_b9f9080338dc.slice/crio-b08534623e7f3b673d2ae117526770d8792ada09d7e9f75b89a33bdb68f8d152 WatchSource:0}: Error finding container b08534623e7f3b673d2ae117526770d8792ada09d7e9f75b89a33bdb68f8d152: Status 404 returned error can't find the container with id b08534623e7f3b673d2ae117526770d8792ada09d7e9f75b89a33bdb68f8d152 Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.120877 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod130c27ae_5c64_4c06_b5e8_9a1aacea630e.slice/crio-fe5a2fc905fcd51f92c8a729a298cf8f6dc3c04492428d5e06828276a3ab4837 WatchSource:0}: Error finding container fe5a2fc905fcd51f92c8a729a298cf8f6dc3c04492428d5e06828276a3ab4837: Status 404 returned error can't find the container with id fe5a2fc905fcd51f92c8a729a298cf8f6dc3c04492428d5e06828276a3ab4837 Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.124300 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cc57759_7dc1_4360_8d81_4744355ec4df.slice/crio-84f8dd6ad14e0dc49ca5dcad6c06769f8f41289f601e4e6d083f46161f82efa4 WatchSource:0}: Error finding container 84f8dd6ad14e0dc49ca5dcad6c06769f8f41289f601e4e6d083f46161f82efa4: Status 404 returned error can't find the container with id 84f8dd6ad14e0dc49ca5dcad6c06769f8f41289f601e4e6d083f46161f82efa4 Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.125404 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd"] Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.134717 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd"] Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.135547 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1558cc4_7188_42b8_858a_e83606895217.slice/crio-9f1e09ce04d130aeb7970ac1eb0cdb1d425510d56fb9a100e23715d38e8c5024 WatchSource:0}: Error finding container 9f1e09ce04d130aeb7970ac1eb0cdb1d425510d56fb9a100e23715d38e8c5024: Status 404 returned error can't find the container with id 9f1e09ce04d130aeb7970ac1eb0cdb1d425510d56fb9a100e23715d38e8c5024 Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.143256 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaabafe79_ea6f_4f01_9efb_e1c1a9e2334e.slice/crio-fa0ce16435af993c4112a6a25fb3156c83e1c6600024986dd83c6b4b3d8992b7 WatchSource:0}: Error finding container fa0ce16435af993c4112a6a25fb3156c83e1c6600024986dd83c6b4b3d8992b7: Status 404 returned error can't find the container with id fa0ce16435af993c4112a6a25fb3156c83e1c6600024986dd83c6b4b3d8992b7 Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.147743 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t"] Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.379274 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7"] Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.407147 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb"] Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.409000 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70440288_e0e0_42fd_b310_da1267956b3d.slice/crio-4734e9fd0eb09f603b66ad9caf230086c6207c16cb89ee08b76786b108046c27 WatchSource:0}: Error finding container 4734e9fd0eb09f603b66ad9caf230086c6207c16cb89ee08b76786b108046c27: Status 404 returned error can't find the container with id 4734e9fd0eb09f603b66ad9caf230086c6207c16cb89ee08b76786b108046c27 Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.413912 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx"] Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.420621 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd"] Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.427061 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9"] Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.445441 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36725987_d3b2_4216_bfe0_69fef2d8bfd7.slice/crio-00d39bdf34e360680e8b9ecd056cacd8c81922a9107a668033ecd46c80073084 WatchSource:0}: Error finding container 00d39bdf34e360680e8b9ecd056cacd8c81922a9107a668033ecd46c80073084: Status 404 returned error can't find the container with id 00d39bdf34e360680e8b9ecd056cacd8c81922a9107a668033ecd46c80073084 Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.457256 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73e35a67_364a_487c_a7bf_5e537cbbecbe.slice/crio-b7d13a91a41f71a44a065c84d0c72c8c537a40e6180545c6ea48b14aed3ca17e WatchSource:0}: Error finding container b7d13a91a41f71a44a065c84d0c72c8c537a40e6180545c6ea48b14aed3ca17e: Status 404 returned error can't find the container with id b7d13a91a41f71a44a065c84d0c72c8c537a40e6180545c6ea48b14aed3ca17e Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.547829 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.548024 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.548117 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert podName:88292fa9-b45e-4681-97df-e05a423895cc nodeName:}" failed. No retries permitted until 2026-01-06 14:54:19.548080041 +0000 UTC m=+1056.175546359 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert") pod "openstack-baremetal-operator-controller-manager-78948ddfd785mfz" (UID: "88292fa9-b45e-4681-97df-e05a423895cc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.739199 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks"] Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.739237 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx"] Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.747137 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h"] Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.753244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.753356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.754262 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.754425 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:19.754309415 +0000 UTC m=+1056.381775733 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "metrics-server-cert" not found Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.755547 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.756068 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:19.755584359 +0000 UTC m=+1056.383050677 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "webhook-server-cert" not found Jan 06 14:54:17 crc kubenswrapper[4744]: I0106 14:54:17.758469 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr"] Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.821182 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfec21671_aa20_4019_bc89_9e0fc135e394.slice/crio-4f09b6ad813dfc89506ee7ec4f53a30c732fe78ed9bcd91b7b95e78757acd934 WatchSource:0}: Error finding container 4f09b6ad813dfc89506ee7ec4f53a30c732fe78ed9bcd91b7b95e78757acd934: Status 404 returned error can't find the container with id 4f09b6ad813dfc89506ee7ec4f53a30c732fe78ed9bcd91b7b95e78757acd934 Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.835224 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ea5e62f_608a_4fa7_8887_691c1c314a6e.slice/crio-08afcfe4b46da5b4c4446bd7c7d9342591a540227d7134f9c055858fbfd04fba WatchSource:0}: Error finding container 08afcfe4b46da5b4c4446bd7c7d9342591a540227d7134f9c055858fbfd04fba: Status 404 returned error can't find the container with id 08afcfe4b46da5b4c4446bd7c7d9342591a540227d7134f9c055858fbfd04fba Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.842219 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:f0ece9a81e4be3dbc1ff752a951970380546d8c0dea910953f862c219444b97a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b4jkc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-9dbdf6486-9j2kx_openstack-operators(1ea5e62f-608a-4fa7-8887-691c1c314a6e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.843649 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" podUID="1ea5e62f-608a-4fa7-8887-691c1c314a6e" Jan 06 14:54:17 crc kubenswrapper[4744]: W0106 14:54:17.847778 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd18584e4_6488_4fc3_9992_9ded578fd05e.slice/crio-c3c7b412551baaa5f150f8351473648564af92f28946e770c95fa26922735b84 WatchSource:0}: Error finding container c3c7b412551baaa5f150f8351473648564af92f28946e770c95fa26922735b84: Status 404 returned error can't find the container with id c3c7b412551baaa5f150f8351473648564af92f28946e770c95fa26922735b84 Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.852947 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.224:5001/openstack-k8s-operators/telemetry-operator:9533fa79d915abe9beaf16e5c08baaa4a197eecd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-shhzp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5649998c8c-g454h_openstack-operators(d18584e4-6488-4fc3-9992-9ded578fd05e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 06 14:54:17 crc kubenswrapper[4744]: E0106 14:54:17.854444 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" podUID="d18584e4-6488-4fc3-9992-9ded578fd05e" Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.030523 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" event={"ID":"70440288-e0e0-42fd-b310-da1267956b3d","Type":"ContainerStarted","Data":"4734e9fd0eb09f603b66ad9caf230086c6207c16cb89ee08b76786b108046c27"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.031915 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" event={"ID":"8cc57759-7dc1-4360-8d81-4744355ec4df","Type":"ContainerStarted","Data":"84f8dd6ad14e0dc49ca5dcad6c06769f8f41289f601e4e6d083f46161f82efa4"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.033676 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" event={"ID":"d18584e4-6488-4fc3-9992-9ded578fd05e","Type":"ContainerStarted","Data":"c3c7b412551baaa5f150f8351473648564af92f28946e770c95fa26922735b84"} Jan 06 14:54:18 crc kubenswrapper[4744]: E0106 14:54:18.034898 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.224:5001/openstack-k8s-operators/telemetry-operator:9533fa79d915abe9beaf16e5c08baaa4a197eecd\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" podUID="d18584e4-6488-4fc3-9992-9ded578fd05e" Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.035566 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" event={"ID":"0abaa053-82ac-4b9f-a4d2-2bc5024a97eb","Type":"ContainerStarted","Data":"b6b50d7d3ceab7dc6456d9a6fa9030cae50638cba0fde0b4382ffc791f30a961"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.036672 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" event={"ID":"1ea5e62f-608a-4fa7-8887-691c1c314a6e","Type":"ContainerStarted","Data":"08afcfe4b46da5b4c4446bd7c7d9342591a540227d7134f9c055858fbfd04fba"} Jan 06 14:54:18 crc kubenswrapper[4744]: E0106 14:54:18.040569 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:f0ece9a81e4be3dbc1ff752a951970380546d8c0dea910953f862c219444b97a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" podUID="1ea5e62f-608a-4fa7-8887-691c1c314a6e" Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.041003 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" event={"ID":"65ec28e9-38c5-4972-b8b8-1447de87ee85","Type":"ContainerStarted","Data":"eef7557fbdad95256a1e04d9a25f2d9e3dda1f128a4a156b9e305d7c61e3ef76"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.043657 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" event={"ID":"f1558cc4-7188-42b8-858a-e83606895217","Type":"ContainerStarted","Data":"9f1e09ce04d130aeb7970ac1eb0cdb1d425510d56fb9a100e23715d38e8c5024"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.046753 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" event={"ID":"73e35a67-364a-487c-a7bf-5e537cbbecbe","Type":"ContainerStarted","Data":"b7d13a91a41f71a44a065c84d0c72c8c537a40e6180545c6ea48b14aed3ca17e"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.054649 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" event={"ID":"96fdaf1c-2733-4272-abda-fb28fa6f6a6a","Type":"ContainerStarted","Data":"2849cc1e533d0077f46f505a01d6bd5a32ab1ba95734d017ed2b7c552b0cf66c"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.058001 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" event={"ID":"36725987-d3b2-4216-bfe0-69fef2d8bfd7","Type":"ContainerStarted","Data":"00d39bdf34e360680e8b9ecd056cacd8c81922a9107a668033ecd46c80073084"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.060136 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" event={"ID":"9ef85238-0ee3-44af-98ad-734994758fec","Type":"ContainerStarted","Data":"b7080aace9b28b0a7392de02577de938df1b161a641f2d328536809935ca4cbb"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.063920 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" event={"ID":"aabafe79-ea6f-4f01-9efb-e1c1a9e2334e","Type":"ContainerStarted","Data":"fa0ce16435af993c4112a6a25fb3156c83e1c6600024986dd83c6b4b3d8992b7"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.083655 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" event={"ID":"5a327dca-6d91-4a3d-a824-b9f9080338dc","Type":"ContainerStarted","Data":"b08534623e7f3b673d2ae117526770d8792ada09d7e9f75b89a33bdb68f8d152"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.087973 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" event={"ID":"130c27ae-5c64-4c06-b5e8-9a1aacea630e","Type":"ContainerStarted","Data":"fe5a2fc905fcd51f92c8a729a298cf8f6dc3c04492428d5e06828276a3ab4837"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.091588 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" event={"ID":"fec21671-aa20-4019-bc89-9e0fc135e394","Type":"ContainerStarted","Data":"4f09b6ad813dfc89506ee7ec4f53a30c732fe78ed9bcd91b7b95e78757acd934"} Jan 06 14:54:18 crc kubenswrapper[4744]: I0106 14:54:18.994806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:18 crc kubenswrapper[4744]: E0106 14:54:18.995012 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:18 crc kubenswrapper[4744]: E0106 14:54:18.995186 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert podName:749a8719-1599-4225-a5f1-165fbe644aca nodeName:}" failed. No retries permitted until 2026-01-06 14:54:22.995154919 +0000 UTC m=+1059.622621237 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert") pod "infra-operator-controller-manager-6d99759cf-pgssl" (UID: "749a8719-1599-4225-a5f1-165fbe644aca") : secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:19 crc kubenswrapper[4744]: E0106 14:54:19.125704 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:f0ece9a81e4be3dbc1ff752a951970380546d8c0dea910953f862c219444b97a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" podUID="1ea5e62f-608a-4fa7-8887-691c1c314a6e" Jan 06 14:54:19 crc kubenswrapper[4744]: E0106 14:54:19.126196 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.224:5001/openstack-k8s-operators/telemetry-operator:9533fa79d915abe9beaf16e5c08baaa4a197eecd\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" podUID="d18584e4-6488-4fc3-9992-9ded578fd05e" Jan 06 14:54:19 crc kubenswrapper[4744]: I0106 14:54:19.619772 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:19 crc kubenswrapper[4744]: E0106 14:54:19.620000 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:19 crc kubenswrapper[4744]: E0106 14:54:19.620301 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert podName:88292fa9-b45e-4681-97df-e05a423895cc nodeName:}" failed. No retries permitted until 2026-01-06 14:54:23.62028517 +0000 UTC m=+1060.247751488 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert") pod "openstack-baremetal-operator-controller-manager-78948ddfd785mfz" (UID: "88292fa9-b45e-4681-97df-e05a423895cc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:19 crc kubenswrapper[4744]: E0106 14:54:19.824953 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 06 14:54:19 crc kubenswrapper[4744]: E0106 14:54:19.825087 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:23.825067056 +0000 UTC m=+1060.452533374 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "metrics-server-cert" not found Jan 06 14:54:19 crc kubenswrapper[4744]: I0106 14:54:19.824249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:19 crc kubenswrapper[4744]: I0106 14:54:19.831635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:19 crc kubenswrapper[4744]: E0106 14:54:19.831785 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 06 14:54:19 crc kubenswrapper[4744]: E0106 14:54:19.831864 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:23.831841995 +0000 UTC m=+1060.459308313 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "webhook-server-cert" not found Jan 06 14:54:23 crc kubenswrapper[4744]: I0106 14:54:23.015925 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:23 crc kubenswrapper[4744]: E0106 14:54:23.016697 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:23 crc kubenswrapper[4744]: E0106 14:54:23.016772 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert podName:749a8719-1599-4225-a5f1-165fbe644aca nodeName:}" failed. No retries permitted until 2026-01-06 14:54:31.016749338 +0000 UTC m=+1067.644215686 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert") pod "infra-operator-controller-manager-6d99759cf-pgssl" (UID: "749a8719-1599-4225-a5f1-165fbe644aca") : secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:23 crc kubenswrapper[4744]: I0106 14:54:23.629690 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:23 crc kubenswrapper[4744]: E0106 14:54:23.629958 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:23 crc kubenswrapper[4744]: E0106 14:54:23.630033 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert podName:88292fa9-b45e-4681-97df-e05a423895cc nodeName:}" failed. No retries permitted until 2026-01-06 14:54:31.630008226 +0000 UTC m=+1068.257474534 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert") pod "openstack-baremetal-operator-controller-manager-78948ddfd785mfz" (UID: "88292fa9-b45e-4681-97df-e05a423895cc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:23 crc kubenswrapper[4744]: I0106 14:54:23.837990 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:23 crc kubenswrapper[4744]: E0106 14:54:23.838449 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 06 14:54:23 crc kubenswrapper[4744]: E0106 14:54:23.838539 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:31.83851763 +0000 UTC m=+1068.465984138 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "webhook-server-cert" not found Jan 06 14:54:23 crc kubenswrapper[4744]: I0106 14:54:23.840619 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:23 crc kubenswrapper[4744]: E0106 14:54:23.841275 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 06 14:54:23 crc kubenswrapper[4744]: E0106 14:54:23.841835 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:31.841811067 +0000 UTC m=+1068.469277385 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "metrics-server-cert" not found Jan 06 14:54:31 crc kubenswrapper[4744]: I0106 14:54:31.073634 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:31 crc kubenswrapper[4744]: E0106 14:54:31.073852 4744 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:31 crc kubenswrapper[4744]: E0106 14:54:31.074452 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert podName:749a8719-1599-4225-a5f1-165fbe644aca nodeName:}" failed. No retries permitted until 2026-01-06 14:54:47.07441941 +0000 UTC m=+1083.701885768 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert") pod "infra-operator-controller-manager-6d99759cf-pgssl" (UID: "749a8719-1599-4225-a5f1-165fbe644aca") : secret "infra-operator-webhook-server-cert" not found Jan 06 14:54:31 crc kubenswrapper[4744]: I0106 14:54:31.688007 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:31 crc kubenswrapper[4744]: E0106 14:54:31.688204 4744 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:31 crc kubenswrapper[4744]: E0106 14:54:31.688273 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert podName:88292fa9-b45e-4681-97df-e05a423895cc nodeName:}" failed. No retries permitted until 2026-01-06 14:54:47.688255882 +0000 UTC m=+1084.315722200 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert") pod "openstack-baremetal-operator-controller-manager-78948ddfd785mfz" (UID: "88292fa9-b45e-4681-97df-e05a423895cc") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 06 14:54:31 crc kubenswrapper[4744]: I0106 14:54:31.896329 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:31 crc kubenswrapper[4744]: E0106 14:54:31.896541 4744 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 06 14:54:31 crc kubenswrapper[4744]: E0106 14:54:31.896618 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:47.896596281 +0000 UTC m=+1084.524062609 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "metrics-server-cert" not found Jan 06 14:54:31 crc kubenswrapper[4744]: I0106 14:54:31.896552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:31 crc kubenswrapper[4744]: E0106 14:54:31.896661 4744 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 06 14:54:31 crc kubenswrapper[4744]: E0106 14:54:31.896724 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs podName:7f9fe9dd-bede-4261-9a8a-d23812cffca2 nodeName:}" failed. No retries permitted until 2026-01-06 14:54:47.896705034 +0000 UTC m=+1084.524171452 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs") pod "openstack-operator-controller-manager-5459d89946-2qz6t" (UID: "7f9fe9dd-bede-4261-9a8a-d23812cffca2") : secret "webhook-server-cert" not found Jan 06 14:54:36 crc kubenswrapper[4744]: E0106 14:54:36.032824 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:c846ab4a49272557884db6b976f979e6b9dce1aa73e5eb7872b4472f44602a1c" Jan 06 14:54:36 crc kubenswrapper[4744]: E0106 14:54:36.033369 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:c846ab4a49272557884db6b976f979e6b9dce1aa73e5eb7872b4472f44602a1c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b6ftb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-598945d5b8-qc2xd_openstack-operators(8cc57759-7dc1-4360-8d81-4744355ec4df): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:36 crc kubenswrapper[4744]: E0106 14:54:36.034455 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" podUID="8cc57759-7dc1-4360-8d81-4744355ec4df" Jan 06 14:54:37 crc kubenswrapper[4744]: E0106 14:54:37.227570 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:c846ab4a49272557884db6b976f979e6b9dce1aa73e5eb7872b4472f44602a1c\\\"\"" pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" podUID="8cc57759-7dc1-4360-8d81-4744355ec4df" Jan 06 14:54:38 crc kubenswrapper[4744]: E0106 14:54:38.142197 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:202756538820b5fa874d07a71ece4f048f41ccca8228d359c8cd25a00e9c0848" Jan 06 14:54:38 crc kubenswrapper[4744]: E0106 14:54:38.142667 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:202756538820b5fa874d07a71ece4f048f41ccca8228d359c8cd25a00e9c0848,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rdkgx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-f99f54bc8-hwxtd_openstack-operators(aabafe79-ea6f-4f01-9efb-e1c1a9e2334e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:38 crc kubenswrapper[4744]: E0106 14:54:38.143935 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" podUID="aabafe79-ea6f-4f01-9efb-e1c1a9e2334e" Jan 06 14:54:38 crc kubenswrapper[4744]: E0106 14:54:38.299938 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:202756538820b5fa874d07a71ece4f048f41ccca8228d359c8cd25a00e9c0848\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" podUID="aabafe79-ea6f-4f01-9efb-e1c1a9e2334e" Jan 06 14:54:38 crc kubenswrapper[4744]: E0106 14:54:38.660059 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Jan 06 14:54:38 crc kubenswrapper[4744]: E0106 14:54:38.660243 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b9m8h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7cd87b778f-bl6hr_openstack-operators(5a327dca-6d91-4a3d-a824-b9f9080338dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:38 crc kubenswrapper[4744]: E0106 14:54:38.661559 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" podUID="5a327dca-6d91-4a3d-a824-b9f9080338dc" Jan 06 14:54:39 crc kubenswrapper[4744]: E0106 14:54:39.305973 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" podUID="5a327dca-6d91-4a3d-a824-b9f9080338dc" Jan 06 14:54:40 crc kubenswrapper[4744]: E0106 14:54:40.898718 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:c10647131e6fa6afeb11ea28e513b60f22dbfbb4ddc3727850b1fe5799890c41" Jan 06 14:54:40 crc kubenswrapper[4744]: E0106 14:54:40.899556 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:c10647131e6fa6afeb11ea28e513b60f22dbfbb4ddc3727850b1fe5799890c41,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ncsk5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-7b88bfc995-k8n6t_openstack-operators(0abaa053-82ac-4b9f-a4d2-2bc5024a97eb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:40 crc kubenswrapper[4744]: E0106 14:54:40.900930 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" podUID="0abaa053-82ac-4b9f-a4d2-2bc5024a97eb" Jan 06 14:54:41 crc kubenswrapper[4744]: E0106 14:54:41.327619 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:c10647131e6fa6afeb11ea28e513b60f22dbfbb4ddc3727850b1fe5799890c41\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" podUID="0abaa053-82ac-4b9f-a4d2-2bc5024a97eb" Jan 06 14:54:41 crc kubenswrapper[4744]: E0106 14:54:41.399113 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:1b684c4ca525a279deee45980140d895e264526c5c7e0a6981d6fae6cbcaa420" Jan 06 14:54:41 crc kubenswrapper[4744]: E0106 14:54:41.399325 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:1b684c4ca525a279deee45980140d895e264526c5c7e0a6981d6fae6cbcaa420,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cjv6x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-9b6f8f78c-fvltb_openstack-operators(9ef85238-0ee3-44af-98ad-734994758fec): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:41 crc kubenswrapper[4744]: E0106 14:54:41.400517 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" podUID="9ef85238-0ee3-44af-98ad-734994758fec" Jan 06 14:54:42 crc kubenswrapper[4744]: E0106 14:54:42.336481 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:1b684c4ca525a279deee45980140d895e264526c5c7e0a6981d6fae6cbcaa420\\\"\"" pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" podUID="9ef85238-0ee3-44af-98ad-734994758fec" Jan 06 14:54:44 crc kubenswrapper[4744]: I0106 14:54:44.424637 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:54:44 crc kubenswrapper[4744]: I0106 14:54:44.424922 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:54:44 crc kubenswrapper[4744]: E0106 14:54:44.699581 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:df69e4193043476bc71d0e06ac8bc7bbd17f7b624d495aae6b7c5e5b40c9e1e7" Jan 06 14:54:44 crc kubenswrapper[4744]: E0106 14:54:44.700051 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:df69e4193043476bc71d0e06ac8bc7bbd17f7b624d495aae6b7c5e5b40c9e1e7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x9f47,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-bb586bbf4-p5kz7_openstack-operators(73e35a67-364a-487c-a7bf-5e537cbbecbe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:44 crc kubenswrapper[4744]: E0106 14:54:44.702838 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" podUID="73e35a67-364a-487c-a7bf-5e537cbbecbe" Jan 06 14:54:45 crc kubenswrapper[4744]: E0106 14:54:45.279106 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Jan 06 14:54:45 crc kubenswrapper[4744]: E0106 14:54:45.279310 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4d5ds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-bf6d4f946-8nnk9_openstack-operators(36725987-d3b2-4216-bfe0-69fef2d8bfd7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:45 crc kubenswrapper[4744]: E0106 14:54:45.280608 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" podUID="36725987-d3b2-4216-bfe0-69fef2d8bfd7" Jan 06 14:54:45 crc kubenswrapper[4744]: E0106 14:54:45.381641 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:df69e4193043476bc71d0e06ac8bc7bbd17f7b624d495aae6b7c5e5b40c9e1e7\\\"\"" pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" podUID="73e35a67-364a-487c-a7bf-5e537cbbecbe" Jan 06 14:54:45 crc kubenswrapper[4744]: E0106 14:54:45.382110 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" podUID="36725987-d3b2-4216-bfe0-69fef2d8bfd7" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.089821 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.111695 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/749a8719-1599-4225-a5f1-165fbe644aca-cert\") pod \"infra-operator-controller-manager-6d99759cf-pgssl\" (UID: \"749a8719-1599-4225-a5f1-165fbe644aca\") " pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:47 crc kubenswrapper[4744]: E0106 14:54:47.155637 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Jan 06 14:54:47 crc kubenswrapper[4744]: E0106 14:54:47.156034 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mlgss,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-66f8b87655-j8rsd_openstack-operators(3653611e-a619-401f-a964-70d245bb4957): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:47 crc kubenswrapper[4744]: E0106 14:54:47.157329 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" podUID="3653611e-a619-401f-a964-70d245bb4957" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.322548 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-bxskh" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.331204 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:54:47 crc kubenswrapper[4744]: E0106 14:54:47.401026 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a\\\"\"" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" podUID="3653611e-a619-401f-a964-70d245bb4957" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.699230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.715841 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/88292fa9-b45e-4681-97df-e05a423895cc-cert\") pod \"openstack-baremetal-operator-controller-manager-78948ddfd785mfz\" (UID: \"88292fa9-b45e-4681-97df-e05a423895cc\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.904076 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.904332 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.919522 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-metrics-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.922786 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/7f9fe9dd-bede-4261-9a8a-d23812cffca2-webhook-certs\") pod \"openstack-operator-controller-manager-5459d89946-2qz6t\" (UID: \"7f9fe9dd-bede-4261-9a8a-d23812cffca2\") " pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:47 crc kubenswrapper[4744]: I0106 14:54:47.992609 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-ldrmw" Jan 06 14:54:48 crc kubenswrapper[4744]: I0106 14:54:48.001446 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:54:48 crc kubenswrapper[4744]: I0106 14:54:48.177529 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-fff9l" Jan 06 14:54:48 crc kubenswrapper[4744]: I0106 14:54:48.185759 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:54:48 crc kubenswrapper[4744]: E0106 14:54:48.858783 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:b7111c690e8fda3cb0c5969bcfa68308907fd0cf05f73ecdcb9ac1423aa7bba3" Jan 06 14:54:48 crc kubenswrapper[4744]: E0106 14:54:48.859325 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:b7111c690e8fda3cb0c5969bcfa68308907fd0cf05f73ecdcb9ac1423aa7bba3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9gsn7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-7f5ddd8d7b-q9k7b_openstack-operators(f1558cc4-7188-42b8-858a-e83606895217): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:48 crc kubenswrapper[4744]: E0106 14:54:48.861632 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" podUID="f1558cc4-7188-42b8-858a-e83606895217" Jan 06 14:54:49 crc kubenswrapper[4744]: E0106 14:54:49.422639 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:b7111c690e8fda3cb0c5969bcfa68308907fd0cf05f73ecdcb9ac1423aa7bba3\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" podUID="f1558cc4-7188-42b8-858a-e83606895217" Jan 06 14:54:53 crc kubenswrapper[4744]: E0106 14:54:53.225011 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:4e3d234c1398039c2593611f7b0fd2a6b284cafb1563e6737876a265b9af42b6" Jan 06 14:54:53 crc kubenswrapper[4744]: E0106 14:54:53.226064 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:4e3d234c1398039c2593611f7b0fd2a6b284cafb1563e6737876a265b9af42b6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-znbg8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-6c866cfdcb-8q8ks_openstack-operators(65ec28e9-38c5-4972-b8b8-1447de87ee85): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:53 crc kubenswrapper[4744]: E0106 14:54:53.227834 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" podUID="65ec28e9-38c5-4972-b8b8-1447de87ee85" Jan 06 14:54:53 crc kubenswrapper[4744]: E0106 14:54:53.451542 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:4e3d234c1398039c2593611f7b0fd2a6b284cafb1563e6737876a265b9af42b6\\\"\"" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" podUID="65ec28e9-38c5-4972-b8b8-1447de87ee85" Jan 06 14:54:53 crc kubenswrapper[4744]: E0106 14:54:53.813307 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:f0ece9a81e4be3dbc1ff752a951970380546d8c0dea910953f862c219444b97a" Jan 06 14:54:53 crc kubenswrapper[4744]: E0106 14:54:53.813483 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:f0ece9a81e4be3dbc1ff752a951970380546d8c0dea910953f862c219444b97a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b4jkc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-9dbdf6486-9j2kx_openstack-operators(1ea5e62f-608a-4fa7-8887-691c1c314a6e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:53 crc kubenswrapper[4744]: E0106 14:54:53.814676 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" podUID="1ea5e62f-608a-4fa7-8887-691c1c314a6e" Jan 06 14:54:55 crc kubenswrapper[4744]: E0106 14:54:55.072857 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Jan 06 14:54:55 crc kubenswrapper[4744]: E0106 14:54:55.073381 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mqlbs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5fbbf8b6cc-twhxx_openstack-operators(70440288-e0e0-42fd-b310-da1267956b3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:55 crc kubenswrapper[4744]: E0106 14:54:55.076407 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" podUID="70440288-e0e0-42fd-b310-da1267956b3d" Jan 06 14:54:55 crc kubenswrapper[4744]: E0106 14:54:55.473873 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" podUID="70440288-e0e0-42fd-b310-da1267956b3d" Jan 06 14:54:55 crc kubenswrapper[4744]: E0106 14:54:55.956914 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Jan 06 14:54:55 crc kubenswrapper[4744]: E0106 14:54:55.957091 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j72f8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-bjtfr_openstack-operators(fec21671-aa20-4019-bc89-9e0fc135e394): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:55 crc kubenswrapper[4744]: E0106 14:54:55.958540 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" podUID="fec21671-aa20-4019-bc89-9e0fc135e394" Jan 06 14:54:56 crc kubenswrapper[4744]: E0106 14:54:56.123953 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.224:5001/openstack-k8s-operators/telemetry-operator:9533fa79d915abe9beaf16e5c08baaa4a197eecd" Jan 06 14:54:56 crc kubenswrapper[4744]: E0106 14:54:56.124013 4744 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.224:5001/openstack-k8s-operators/telemetry-operator:9533fa79d915abe9beaf16e5c08baaa4a197eecd" Jan 06 14:54:56 crc kubenswrapper[4744]: E0106 14:54:56.124235 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.224:5001/openstack-k8s-operators/telemetry-operator:9533fa79d915abe9beaf16e5c08baaa4a197eecd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-shhzp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5649998c8c-g454h_openstack-operators(d18584e4-6488-4fc3-9992-9ded578fd05e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:54:56 crc kubenswrapper[4744]: E0106 14:54:56.125394 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" podUID="d18584e4-6488-4fc3-9992-9ded578fd05e" Jan 06 14:54:56 crc kubenswrapper[4744]: E0106 14:54:56.485529 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" podUID="fec21671-aa20-4019-bc89-9e0fc135e394" Jan 06 14:54:56 crc kubenswrapper[4744]: I0106 14:54:56.670624 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl"] Jan 06 14:54:56 crc kubenswrapper[4744]: I0106 14:54:56.694562 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t"] Jan 06 14:54:56 crc kubenswrapper[4744]: W0106 14:54:56.742325 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f9fe9dd_bede_4261_9a8a_d23812cffca2.slice/crio-d871bf940d21ec183bb94baa7601a77d6e2c01bb37bdcce647f97b3ce94e5a30 WatchSource:0}: Error finding container d871bf940d21ec183bb94baa7601a77d6e2c01bb37bdcce647f97b3ce94e5a30: Status 404 returned error can't find the container with id d871bf940d21ec183bb94baa7601a77d6e2c01bb37bdcce647f97b3ce94e5a30 Jan 06 14:54:56 crc kubenswrapper[4744]: I0106 14:54:56.819652 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz"] Jan 06 14:54:56 crc kubenswrapper[4744]: W0106 14:54:56.825616 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88292fa9_b45e_4681_97df_e05a423895cc.slice/crio-25f100d3430296d626d75100ff975068d7c5a8c1de045b7ed578e70973c6a8b5 WatchSource:0}: Error finding container 25f100d3430296d626d75100ff975068d7c5a8c1de045b7ed578e70973c6a8b5: Status 404 returned error can't find the container with id 25f100d3430296d626d75100ff975068d7c5a8c1de045b7ed578e70973c6a8b5 Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.490433 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" event={"ID":"aabafe79-ea6f-4f01-9efb-e1c1a9e2334e","Type":"ContainerStarted","Data":"85827422ae9bbee2e921b69796830eabc5c3378fa3fd397acb84e114fa7c82cb"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.490664 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.491968 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" event={"ID":"8035e88d-2a38-4e95-b204-0dc00fc57bfc","Type":"ContainerStarted","Data":"f24010e45c22f201e3450d301902bc0886ba93ddcba5afdc4252eeabfb21ce47"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.492119 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.492968 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" event={"ID":"7f9fe9dd-bede-4261-9a8a-d23812cffca2","Type":"ContainerStarted","Data":"d871bf940d21ec183bb94baa7601a77d6e2c01bb37bdcce647f97b3ce94e5a30"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.494670 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" event={"ID":"9ef85238-0ee3-44af-98ad-734994758fec","Type":"ContainerStarted","Data":"f912ceb834e9a0c156949079e63f7e5c95dde16282e69a46c48be0bf1cd878e3"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.495019 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.496280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" event={"ID":"8421b8a1-9929-43ab-b203-9899c52d2d41","Type":"ContainerStarted","Data":"811e18fda4cb3f266fd62417c1c2348eaa5279435257fe24bf517d3fc4ee7603"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.496639 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.497845 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" event={"ID":"5b589b5b-5f62-4e3d-808b-8c404d1432ac","Type":"ContainerStarted","Data":"da737d97ee83bb9f4551dfc2579f70e70c4276a9e99828cee767c1c0fd0f5ee2"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.498946 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.499931 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" event={"ID":"88292fa9-b45e-4681-97df-e05a423895cc","Type":"ContainerStarted","Data":"25f100d3430296d626d75100ff975068d7c5a8c1de045b7ed578e70973c6a8b5"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.501132 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" event={"ID":"8cc57759-7dc1-4360-8d81-4744355ec4df","Type":"ContainerStarted","Data":"e72f1610657294654536727551173deaad1894faa373f3b22b2baadabdbc500d"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.501517 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.502749 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" event={"ID":"5a327dca-6d91-4a3d-a824-b9f9080338dc","Type":"ContainerStarted","Data":"a160fb0120699d286f95aa93906aa44b298f113e1e6578de74e45f755a77b6d9"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.503092 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.504437 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" event={"ID":"96fdaf1c-2733-4272-abda-fb28fa6f6a6a","Type":"ContainerStarted","Data":"4eca21f5545971cf6046a7c7686b88a0c28ac9b402ca9b15a4452393007cf8c4"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.504820 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.506578 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" event={"ID":"130c27ae-5c64-4c06-b5e8-9a1aacea630e","Type":"ContainerStarted","Data":"99bfc9b23a9e01c6b529af3c4e142afce03a2e8ea45ca0378c20ff0c1ef5b5e8"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.506601 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.508082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" event={"ID":"0abaa053-82ac-4b9f-a4d2-2bc5024a97eb","Type":"ContainerStarted","Data":"749aece1e9e1d43f5fa8dceb6d6860ef194d5b65704a3def02ea63d71a376250"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.508301 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.509077 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" event={"ID":"749a8719-1599-4225-a5f1-165fbe644aca","Type":"ContainerStarted","Data":"d3841a1db0a1773c3e5f91fcd7bfd471c5d9ee46755cd87783bd03b55864f728"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.510406 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" event={"ID":"66470ebb-8a02-4745-941b-5b6a1b97df1c","Type":"ContainerStarted","Data":"2cd4e40453030981deddc6976ba94a4405252f3c54a4c45af011dd90e06df9bb"} Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.510546 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.520051 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" podStartSLOduration=3.450252358 podStartE2EDuration="42.520033191s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.149818309 +0000 UTC m=+1053.777284637" lastFinishedPulling="2026-01-06 14:54:56.219599112 +0000 UTC m=+1092.847065470" observedRunningTime="2026-01-06 14:54:57.514697259 +0000 UTC m=+1094.142163577" watchObservedRunningTime="2026-01-06 14:54:57.520033191 +0000 UTC m=+1094.147499509" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.539990 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" podStartSLOduration=6.86062188 podStartE2EDuration="43.539974058s" podCreationTimestamp="2026-01-06 14:54:14 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.123210716 +0000 UTC m=+1053.750677024" lastFinishedPulling="2026-01-06 14:54:53.802562884 +0000 UTC m=+1090.430029202" observedRunningTime="2026-01-06 14:54:57.537436931 +0000 UTC m=+1094.164903259" watchObservedRunningTime="2026-01-06 14:54:57.539974058 +0000 UTC m=+1094.167440376" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.559589 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" podStartSLOduration=3.240966045 podStartE2EDuration="42.559574606s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.113873959 +0000 UTC m=+1053.741340277" lastFinishedPulling="2026-01-06 14:54:56.43248252 +0000 UTC m=+1093.059948838" observedRunningTime="2026-01-06 14:54:57.553386423 +0000 UTC m=+1094.180852741" watchObservedRunningTime="2026-01-06 14:54:57.559574606 +0000 UTC m=+1094.187040914" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.589655 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" podStartSLOduration=4.998285105 podStartE2EDuration="42.589641591s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.464026178 +0000 UTC m=+1054.091492486" lastFinishedPulling="2026-01-06 14:54:55.055382614 +0000 UTC m=+1091.682848972" observedRunningTime="2026-01-06 14:54:57.584600028 +0000 UTC m=+1094.212066346" watchObservedRunningTime="2026-01-06 14:54:57.589641591 +0000 UTC m=+1094.217107909" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.607962 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" podStartSLOduration=4.136323602 podStartE2EDuration="42.607944615s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:16.583635798 +0000 UTC m=+1053.211102116" lastFinishedPulling="2026-01-06 14:54:55.055256811 +0000 UTC m=+1091.682723129" observedRunningTime="2026-01-06 14:54:57.601985598 +0000 UTC m=+1094.229451916" watchObservedRunningTime="2026-01-06 14:54:57.607944615 +0000 UTC m=+1094.235410933" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.619173 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" podStartSLOduration=3.615667603 podStartE2EDuration="42.619141531s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.429012392 +0000 UTC m=+1054.056478710" lastFinishedPulling="2026-01-06 14:54:56.4324863 +0000 UTC m=+1093.059952638" observedRunningTime="2026-01-06 14:54:57.616544733 +0000 UTC m=+1094.244011071" watchObservedRunningTime="2026-01-06 14:54:57.619141531 +0000 UTC m=+1094.246607849" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.631791 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" podStartSLOduration=3.535048591 podStartE2EDuration="42.631776046s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.125979509 +0000 UTC m=+1053.753445827" lastFinishedPulling="2026-01-06 14:54:56.222706914 +0000 UTC m=+1092.850173282" observedRunningTime="2026-01-06 14:54:57.631153999 +0000 UTC m=+1094.258620317" watchObservedRunningTime="2026-01-06 14:54:57.631776046 +0000 UTC m=+1094.259242364" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.647515 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" podStartSLOduration=3.512703499 podStartE2EDuration="42.647498191s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.118503551 +0000 UTC m=+1053.745969869" lastFinishedPulling="2026-01-06 14:54:56.253298233 +0000 UTC m=+1092.880764561" observedRunningTime="2026-01-06 14:54:57.643876736 +0000 UTC m=+1094.271343054" watchObservedRunningTime="2026-01-06 14:54:57.647498191 +0000 UTC m=+1094.274964509" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.663034 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" podStartSLOduration=4.137244667 podStartE2EDuration="42.663019182s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:16.583678359 +0000 UTC m=+1053.211144677" lastFinishedPulling="2026-01-06 14:54:55.109452874 +0000 UTC m=+1091.736919192" observedRunningTime="2026-01-06 14:54:57.659304424 +0000 UTC m=+1094.286770742" watchObservedRunningTime="2026-01-06 14:54:57.663019182 +0000 UTC m=+1094.290485500" Jan 06 14:54:57 crc kubenswrapper[4744]: I0106 14:54:57.695968 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" podStartSLOduration=6.641481056 podStartE2EDuration="43.695951963s" podCreationTimestamp="2026-01-06 14:54:14 +0000 UTC" firstStartedPulling="2026-01-06 14:54:16.150644447 +0000 UTC m=+1052.778110765" lastFinishedPulling="2026-01-06 14:54:53.205115354 +0000 UTC m=+1089.832581672" observedRunningTime="2026-01-06 14:54:57.687942261 +0000 UTC m=+1094.315408569" watchObservedRunningTime="2026-01-06 14:54:57.695951963 +0000 UTC m=+1094.323418271" Jan 06 14:54:58 crc kubenswrapper[4744]: I0106 14:54:58.730750 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" podStartSLOduration=7.10923885 podStartE2EDuration="43.730730417s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:16.583601127 +0000 UTC m=+1053.211067445" lastFinishedPulling="2026-01-06 14:54:53.205092664 +0000 UTC m=+1089.832559012" observedRunningTime="2026-01-06 14:54:57.710468977 +0000 UTC m=+1094.337935295" watchObservedRunningTime="2026-01-06 14:54:58.730730417 +0000 UTC m=+1095.358196735" Jan 06 14:54:59 crc kubenswrapper[4744]: I0106 14:54:59.533476 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" event={"ID":"7f9fe9dd-bede-4261-9a8a-d23812cffca2","Type":"ContainerStarted","Data":"4bfb136ef2fb0bc4fbf423a6a474339169dd2e43c7b9def8e6e2efb8a988e27f"} Jan 06 14:54:59 crc kubenswrapper[4744]: I0106 14:54:59.572480 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" podStartSLOduration=44.572458637 podStartE2EDuration="44.572458637s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:54:59.560632524 +0000 UTC m=+1096.188098852" watchObservedRunningTime="2026-01-06 14:54:59.572458637 +0000 UTC m=+1096.199924955" Jan 06 14:55:00 crc kubenswrapper[4744]: I0106 14:55:00.543382 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" event={"ID":"73e35a67-364a-487c-a7bf-5e537cbbecbe","Type":"ContainerStarted","Data":"b46b8805e7c978529c133bef6dc4cd5cade982f9c2ea5bc93f2a3c3eaa22c607"} Jan 06 14:55:00 crc kubenswrapper[4744]: I0106 14:55:00.544282 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" Jan 06 14:55:00 crc kubenswrapper[4744]: I0106 14:55:00.545593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" event={"ID":"3653611e-a619-401f-a964-70d245bb4957","Type":"ContainerStarted","Data":"f7c9cbe9471b682ba6232577abf6340e6fedcbd0a9d457a95e9c7082021a8551"} Jan 06 14:55:00 crc kubenswrapper[4744]: I0106 14:55:00.545729 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" Jan 06 14:55:00 crc kubenswrapper[4744]: I0106 14:55:00.548612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" event={"ID":"36725987-d3b2-4216-bfe0-69fef2d8bfd7","Type":"ContainerStarted","Data":"aea0ddf2a7ecd676aaccf26ab5e34ace6bf0f0f0ae9cf96e9fec50334310fd0a"} Jan 06 14:55:00 crc kubenswrapper[4744]: I0106 14:55:00.548744 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:55:00 crc kubenswrapper[4744]: I0106 14:55:00.571725 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" podStartSLOduration=3.237480221 podStartE2EDuration="45.57169733s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.463388872 +0000 UTC m=+1054.090855190" lastFinishedPulling="2026-01-06 14:54:59.797605971 +0000 UTC m=+1096.425072299" observedRunningTime="2026-01-06 14:55:00.559835267 +0000 UTC m=+1097.187301605" watchObservedRunningTime="2026-01-06 14:55:00.57169733 +0000 UTC m=+1097.199163648" Jan 06 14:55:00 crc kubenswrapper[4744]: I0106 14:55:00.583970 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" podStartSLOduration=3.549001439 podStartE2EDuration="45.583949894s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.45955426 +0000 UTC m=+1054.087020578" lastFinishedPulling="2026-01-06 14:54:59.494502705 +0000 UTC m=+1096.121969033" observedRunningTime="2026-01-06 14:55:00.582948108 +0000 UTC m=+1097.210414426" watchObservedRunningTime="2026-01-06 14:55:00.583949894 +0000 UTC m=+1097.211416212" Jan 06 14:55:00 crc kubenswrapper[4744]: I0106 14:55:00.604093 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" podStartSLOduration=2.670135035 podStartE2EDuration="46.604068576s" podCreationTimestamp="2026-01-06 14:54:14 +0000 UTC" firstStartedPulling="2026-01-06 14:54:16.281848367 +0000 UTC m=+1052.909314685" lastFinishedPulling="2026-01-06 14:55:00.215781908 +0000 UTC m=+1096.843248226" observedRunningTime="2026-01-06 14:55:00.599824264 +0000 UTC m=+1097.227290582" watchObservedRunningTime="2026-01-06 14:55:00.604068576 +0000 UTC m=+1097.231534894" Jan 06 14:55:03 crc kubenswrapper[4744]: I0106 14:55:03.574202 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" event={"ID":"f1558cc4-7188-42b8-858a-e83606895217","Type":"ContainerStarted","Data":"0d622f15ebc4fc77de68c3b283893b32a2f4dccf97fed62c02877010237bc51c"} Jan 06 14:55:03 crc kubenswrapper[4744]: I0106 14:55:03.574807 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" Jan 06 14:55:03 crc kubenswrapper[4744]: I0106 14:55:03.580595 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" event={"ID":"749a8719-1599-4225-a5f1-165fbe644aca","Type":"ContainerStarted","Data":"d62697158c568e0978e0dcf5a4a8d3748b895049bd29b0b3f75a4b7cdb16418e"} Jan 06 14:55:03 crc kubenswrapper[4744]: I0106 14:55:03.580781 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:55:03 crc kubenswrapper[4744]: I0106 14:55:03.582935 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" event={"ID":"88292fa9-b45e-4681-97df-e05a423895cc","Type":"ContainerStarted","Data":"3b23462e34e117200d14969b4493cc13fd4a0421f75ae84e603dccbcf18203cc"} Jan 06 14:55:03 crc kubenswrapper[4744]: I0106 14:55:03.583266 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:55:03 crc kubenswrapper[4744]: I0106 14:55:03.595453 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" podStartSLOduration=3.068777681 podStartE2EDuration="48.595432662s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.137551735 +0000 UTC m=+1053.765018053" lastFinishedPulling="2026-01-06 14:55:02.664206726 +0000 UTC m=+1099.291673034" observedRunningTime="2026-01-06 14:55:03.593330726 +0000 UTC m=+1100.220797054" watchObservedRunningTime="2026-01-06 14:55:03.595432662 +0000 UTC m=+1100.222898980" Jan 06 14:55:03 crc kubenswrapper[4744]: I0106 14:55:03.618466 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" podStartSLOduration=42.784531064 podStartE2EDuration="48.61844806s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:56.828795981 +0000 UTC m=+1093.456262299" lastFinishedPulling="2026-01-06 14:55:02.662712977 +0000 UTC m=+1099.290179295" observedRunningTime="2026-01-06 14:55:03.617954937 +0000 UTC m=+1100.245421255" watchObservedRunningTime="2026-01-06 14:55:03.61844806 +0000 UTC m=+1100.245914368" Jan 06 14:55:03 crc kubenswrapper[4744]: I0106 14:55:03.635137 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" podStartSLOduration=42.665417335 podStartE2EDuration="48.635119951s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:56.695981789 +0000 UTC m=+1093.323448107" lastFinishedPulling="2026-01-06 14:55:02.665684405 +0000 UTC m=+1099.293150723" observedRunningTime="2026-01-06 14:55:03.632559543 +0000 UTC m=+1100.260025861" watchObservedRunningTime="2026-01-06 14:55:03.635119951 +0000 UTC m=+1100.262586269" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.326898 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-78979fc445-bm4m2" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.354293 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66f8b87655-j8rsd" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.398672 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7b549fc966-clgkb" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.441815 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-658dd65b86-pnxwz" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.591108 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-568985c78-dhg75" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.601995 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-f99f54bc8-hwxtd" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.615734 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-598945d5b8-qc2xd" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.624438 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-f6f74d6db-lhppk" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.654813 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" Jan 06 14:55:05 crc kubenswrapper[4744]: E0106 14:55:05.715607 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:f0ece9a81e4be3dbc1ff752a951970380546d8c0dea910953f862c219444b97a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" podUID="1ea5e62f-608a-4fa7-8887-691c1c314a6e" Jan 06 14:55:05 crc kubenswrapper[4744]: I0106 14:55:05.979975 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7cd87b778f-bl6hr" Jan 06 14:55:06 crc kubenswrapper[4744]: I0106 14:55:06.107646 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-68c649d9d-dzwnd" Jan 06 14:55:06 crc kubenswrapper[4744]: I0106 14:55:06.145147 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" Jan 06 14:55:06 crc kubenswrapper[4744]: I0106 14:55:06.154037 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bf6d4f946-8nnk9" Jan 06 14:55:06 crc kubenswrapper[4744]: I0106 14:55:06.180750 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-9b6f8f78c-fvltb" Jan 06 14:55:06 crc kubenswrapper[4744]: I0106 14:55:06.223822 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-bb586bbf4-p5kz7" Jan 06 14:55:06 crc kubenswrapper[4744]: I0106 14:55:06.623097 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" event={"ID":"65ec28e9-38c5-4972-b8b8-1447de87ee85","Type":"ContainerStarted","Data":"d6ecc71f25189774d15ce1fff660c943f09288a8d54d823f4eda2d6cbf15ad94"} Jan 06 14:55:06 crc kubenswrapper[4744]: I0106 14:55:06.626086 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" Jan 06 14:55:06 crc kubenswrapper[4744]: I0106 14:55:06.654359 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" podStartSLOduration=3.229506521 podStartE2EDuration="51.654339404s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.793392548 +0000 UTC m=+1054.420858866" lastFinishedPulling="2026-01-06 14:55:06.218225431 +0000 UTC m=+1102.845691749" observedRunningTime="2026-01-06 14:55:06.644216886 +0000 UTC m=+1103.271683214" watchObservedRunningTime="2026-01-06 14:55:06.654339404 +0000 UTC m=+1103.281805722" Jan 06 14:55:07 crc kubenswrapper[4744]: I0106 14:55:07.338750 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6d99759cf-pgssl" Jan 06 14:55:08 crc kubenswrapper[4744]: I0106 14:55:08.009311 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-78948ddfd785mfz" Jan 06 14:55:08 crc kubenswrapper[4744]: I0106 14:55:08.192403 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5459d89946-2qz6t" Jan 06 14:55:09 crc kubenswrapper[4744]: I0106 14:55:09.646438 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" event={"ID":"70440288-e0e0-42fd-b310-da1267956b3d","Type":"ContainerStarted","Data":"419b949ae92892a4b0f586e74493a65fab5bf54943f46f12fbea9ed2f2df19e7"} Jan 06 14:55:09 crc kubenswrapper[4744]: I0106 14:55:09.647413 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" Jan 06 14:55:09 crc kubenswrapper[4744]: I0106 14:55:09.660712 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" podStartSLOduration=2.927422613 podStartE2EDuration="54.660694565s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.448986041 +0000 UTC m=+1054.076452359" lastFinishedPulling="2026-01-06 14:55:09.182257993 +0000 UTC m=+1105.809724311" observedRunningTime="2026-01-06 14:55:09.659239707 +0000 UTC m=+1106.286706015" watchObservedRunningTime="2026-01-06 14:55:09.660694565 +0000 UTC m=+1106.288160883" Jan 06 14:55:10 crc kubenswrapper[4744]: I0106 14:55:10.662670 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" event={"ID":"fec21671-aa20-4019-bc89-9e0fc135e394","Type":"ContainerStarted","Data":"e057efb34faf0f409f4d972708809c164da2eb92b9d97cdcd531f209930e72bd"} Jan 06 14:55:10 crc kubenswrapper[4744]: I0106 14:55:10.692878 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-bjtfr" podStartSLOduration=3.330748718 podStartE2EDuration="55.69285508s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.827310555 +0000 UTC m=+1054.454776863" lastFinishedPulling="2026-01-06 14:55:10.189416907 +0000 UTC m=+1106.816883225" observedRunningTime="2026-01-06 14:55:10.685191157 +0000 UTC m=+1107.312657475" watchObservedRunningTime="2026-01-06 14:55:10.69285508 +0000 UTC m=+1107.320321438" Jan 06 14:55:10 crc kubenswrapper[4744]: E0106 14:55:10.712559 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.224:5001/openstack-k8s-operators/telemetry-operator:9533fa79d915abe9beaf16e5c08baaa4a197eecd\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" podUID="d18584e4-6488-4fc3-9992-9ded578fd05e" Jan 06 14:55:14 crc kubenswrapper[4744]: I0106 14:55:14.423507 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:55:14 crc kubenswrapper[4744]: I0106 14:55:14.423975 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:55:14 crc kubenswrapper[4744]: I0106 14:55:14.424026 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:55:14 crc kubenswrapper[4744]: I0106 14:55:14.424869 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b95bf86d9d16dc50821869dcd40536e357bac00adf9429e476ea048ffe56203"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 14:55:14 crc kubenswrapper[4744]: I0106 14:55:14.425006 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://7b95bf86d9d16dc50821869dcd40536e357bac00adf9429e476ea048ffe56203" gracePeriod=600 Jan 06 14:55:14 crc kubenswrapper[4744]: I0106 14:55:14.704576 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="7b95bf86d9d16dc50821869dcd40536e357bac00adf9429e476ea048ffe56203" exitCode=0 Jan 06 14:55:14 crc kubenswrapper[4744]: I0106 14:55:14.704650 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"7b95bf86d9d16dc50821869dcd40536e357bac00adf9429e476ea048ffe56203"} Jan 06 14:55:14 crc kubenswrapper[4744]: I0106 14:55:14.704980 4744 scope.go:117] "RemoveContainer" containerID="b7fdf0a873dd4feb170b2380a6eb8f3f910485e646fb343c689eab9c501fa171" Jan 06 14:55:15 crc kubenswrapper[4744]: I0106 14:55:15.725484 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"8cc1261767d494e24246d9d799ed9782619069fb6a47b29f76fd6dc783322bbb"} Jan 06 14:55:15 crc kubenswrapper[4744]: I0106 14:55:15.780042 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7f5ddd8d7b-q9k7b" Jan 06 14:55:16 crc kubenswrapper[4744]: I0106 14:55:16.035306 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5fbbf8b6cc-twhxx" Jan 06 14:55:16 crc kubenswrapper[4744]: I0106 14:55:16.264698 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" Jan 06 14:55:19 crc kubenswrapper[4744]: I0106 14:55:19.756200 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" event={"ID":"1ea5e62f-608a-4fa7-8887-691c1c314a6e","Type":"ContainerStarted","Data":"ed990bc52d8e14f2895151f171ed3f2f7966670bdf981cce4e564410a9a25de1"} Jan 06 14:55:19 crc kubenswrapper[4744]: I0106 14:55:19.756932 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" Jan 06 14:55:19 crc kubenswrapper[4744]: I0106 14:55:19.784114 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" podStartSLOduration=3.827780403 podStartE2EDuration="1m4.784083764s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.842042475 +0000 UTC m=+1054.469508793" lastFinishedPulling="2026-01-06 14:55:18.798345836 +0000 UTC m=+1115.425812154" observedRunningTime="2026-01-06 14:55:19.773487294 +0000 UTC m=+1116.400953652" watchObservedRunningTime="2026-01-06 14:55:19.784083764 +0000 UTC m=+1116.411550122" Jan 06 14:55:22 crc kubenswrapper[4744]: I0106 14:55:22.813634 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" event={"ID":"d18584e4-6488-4fc3-9992-9ded578fd05e","Type":"ContainerStarted","Data":"e015ae67b5feadf891e0ee4b7a587269c21f1081828a846a5d158577405bf35d"} Jan 06 14:55:22 crc kubenswrapper[4744]: I0106 14:55:22.814353 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" Jan 06 14:55:22 crc kubenswrapper[4744]: I0106 14:55:22.845516 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" podStartSLOduration=3.901226254 podStartE2EDuration="1m7.845491091s" podCreationTimestamp="2026-01-06 14:54:15 +0000 UTC" firstStartedPulling="2026-01-06 14:54:17.85282175 +0000 UTC m=+1054.480288068" lastFinishedPulling="2026-01-06 14:55:21.797086557 +0000 UTC m=+1118.424552905" observedRunningTime="2026-01-06 14:55:22.835815536 +0000 UTC m=+1119.463281864" watchObservedRunningTime="2026-01-06 14:55:22.845491091 +0000 UTC m=+1119.472957429" Jan 06 14:55:26 crc kubenswrapper[4744]: I0106 14:55:26.279576 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-9dbdf6486-9j2kx" Jan 06 14:55:36 crc kubenswrapper[4744]: I0106 14:55:36.242074 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" Jan 06 14:55:53 crc kubenswrapper[4744]: I0106 14:55:53.885257 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tmvg4"] Jan 06 14:55:53 crc kubenswrapper[4744]: I0106 14:55:53.887244 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:55:53 crc kubenswrapper[4744]: I0106 14:55:53.899846 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-r2d5g" Jan 06 14:55:53 crc kubenswrapper[4744]: I0106 14:55:53.900138 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Jan 06 14:55:53 crc kubenswrapper[4744]: I0106 14:55:53.900342 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Jan 06 14:55:53 crc kubenswrapper[4744]: I0106 14:55:53.900470 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Jan 06 14:55:53 crc kubenswrapper[4744]: I0106 14:55:53.936444 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tmvg4"] Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.015855 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-config\") pod \"dnsmasq-dns-675f4bcbfc-tmvg4\" (UID: \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.016117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjwgm\" (UniqueName: \"kubernetes.io/projected/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-kube-api-access-qjwgm\") pod \"dnsmasq-dns-675f4bcbfc-tmvg4\" (UID: \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.021410 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tqpfk"] Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.025150 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.029335 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tqpfk"] Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.029572 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.117416 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjwgm\" (UniqueName: \"kubernetes.io/projected/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-kube-api-access-qjwgm\") pod \"dnsmasq-dns-675f4bcbfc-tmvg4\" (UID: \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.117531 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-config\") pod \"dnsmasq-dns-675f4bcbfc-tmvg4\" (UID: \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.118418 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-config\") pod \"dnsmasq-dns-675f4bcbfc-tmvg4\" (UID: \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.139030 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjwgm\" (UniqueName: \"kubernetes.io/projected/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-kube-api-access-qjwgm\") pod \"dnsmasq-dns-675f4bcbfc-tmvg4\" (UID: \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.218560 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tqpfk\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.218620 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-config\") pod \"dnsmasq-dns-78dd6ddcc-tqpfk\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.218716 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c67bp\" (UniqueName: \"kubernetes.io/projected/a1ab7c26-7c69-43d2-a237-085cd33418c2-kube-api-access-c67bp\") pod \"dnsmasq-dns-78dd6ddcc-tqpfk\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.261459 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.320641 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tqpfk\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.320739 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-config\") pod \"dnsmasq-dns-78dd6ddcc-tqpfk\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.320895 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c67bp\" (UniqueName: \"kubernetes.io/projected/a1ab7c26-7c69-43d2-a237-085cd33418c2-kube-api-access-c67bp\") pod \"dnsmasq-dns-78dd6ddcc-tqpfk\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.321775 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tqpfk\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.321775 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-config\") pod \"dnsmasq-dns-78dd6ddcc-tqpfk\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.336736 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c67bp\" (UniqueName: \"kubernetes.io/projected/a1ab7c26-7c69-43d2-a237-085cd33418c2-kube-api-access-c67bp\") pod \"dnsmasq-dns-78dd6ddcc-tqpfk\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.349826 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.698276 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tmvg4"] Jan 06 14:55:54 crc kubenswrapper[4744]: I0106 14:55:54.859407 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tqpfk"] Jan 06 14:55:54 crc kubenswrapper[4744]: W0106 14:55:54.860439 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1ab7c26_7c69_43d2_a237_085cd33418c2.slice/crio-aab4c6a3bfafba9980dfac788f7537d996199f0ab510eb0eb31d543f46fb0a57 WatchSource:0}: Error finding container aab4c6a3bfafba9980dfac788f7537d996199f0ab510eb0eb31d543f46fb0a57: Status 404 returned error can't find the container with id aab4c6a3bfafba9980dfac788f7537d996199f0ab510eb0eb31d543f46fb0a57 Jan 06 14:55:55 crc kubenswrapper[4744]: I0106 14:55:55.126019 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" event={"ID":"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee","Type":"ContainerStarted","Data":"138a5a73da26dbf2bbd8765e595917658a85f391fbc6df87f89f426ceecdf3e9"} Jan 06 14:55:55 crc kubenswrapper[4744]: I0106 14:55:55.127751 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" event={"ID":"a1ab7c26-7c69-43d2-a237-085cd33418c2","Type":"ContainerStarted","Data":"aab4c6a3bfafba9980dfac788f7537d996199f0ab510eb0eb31d543f46fb0a57"} Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.829738 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tmvg4"] Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.855028 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mlnxx"] Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.856483 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.868476 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mlnxx"] Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.877862 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-dns-svc\") pod \"dnsmasq-dns-666b6646f7-mlnxx\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.877891 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tv7b\" (UniqueName: \"kubernetes.io/projected/0d545816-db6b-4953-99a2-716ed95439ac-kube-api-access-4tv7b\") pod \"dnsmasq-dns-666b6646f7-mlnxx\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.877993 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-config\") pod \"dnsmasq-dns-666b6646f7-mlnxx\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.979170 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-config\") pod \"dnsmasq-dns-666b6646f7-mlnxx\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.979243 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-dns-svc\") pod \"dnsmasq-dns-666b6646f7-mlnxx\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.979270 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tv7b\" (UniqueName: \"kubernetes.io/projected/0d545816-db6b-4953-99a2-716ed95439ac-kube-api-access-4tv7b\") pod \"dnsmasq-dns-666b6646f7-mlnxx\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.980858 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-config\") pod \"dnsmasq-dns-666b6646f7-mlnxx\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:56 crc kubenswrapper[4744]: I0106 14:55:56.981365 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-dns-svc\") pod \"dnsmasq-dns-666b6646f7-mlnxx\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.016926 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tv7b\" (UniqueName: \"kubernetes.io/projected/0d545816-db6b-4953-99a2-716ed95439ac-kube-api-access-4tv7b\") pod \"dnsmasq-dns-666b6646f7-mlnxx\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.175424 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tqpfk"] Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.183262 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.193232 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cmsvx"] Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.194874 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.226220 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cmsvx"] Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.287817 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sp5b\" (UniqueName: \"kubernetes.io/projected/aa886d28-0827-4dae-84bf-a69f196a8933-kube-api-access-4sp5b\") pod \"dnsmasq-dns-57d769cc4f-cmsvx\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.287874 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cmsvx\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.287936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-config\") pod \"dnsmasq-dns-57d769cc4f-cmsvx\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.390286 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sp5b\" (UniqueName: \"kubernetes.io/projected/aa886d28-0827-4dae-84bf-a69f196a8933-kube-api-access-4sp5b\") pod \"dnsmasq-dns-57d769cc4f-cmsvx\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.390725 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cmsvx\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.390769 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-config\") pod \"dnsmasq-dns-57d769cc4f-cmsvx\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.391825 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-config\") pod \"dnsmasq-dns-57d769cc4f-cmsvx\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.391885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cmsvx\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.421489 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sp5b\" (UniqueName: \"kubernetes.io/projected/aa886d28-0827-4dae-84bf-a69f196a8933-kube-api-access-4sp5b\") pod \"dnsmasq-dns-57d769cc4f-cmsvx\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.528330 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.805009 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mlnxx"] Jan 06 14:55:57 crc kubenswrapper[4744]: W0106 14:55:57.820753 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d545816_db6b_4953_99a2_716ed95439ac.slice/crio-d19600dbcdd675a8553b7c7817e8bca4a485c84e3a9d6b450b28326ef2ecc821 WatchSource:0}: Error finding container d19600dbcdd675a8553b7c7817e8bca4a485c84e3a9d6b450b28326ef2ecc821: Status 404 returned error can't find the container with id d19600dbcdd675a8553b7c7817e8bca4a485c84e3a9d6b450b28326ef2ecc821 Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.993616 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 06 14:55:57 crc kubenswrapper[4744]: I0106 14:55:57.995008 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:57.998564 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:57.998819 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:57.999517 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:57.999674 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:57.999845 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.000009 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.001758 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-bdxlz" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.025419 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cmsvx"] Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.059288 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.080407 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.082964 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.094858 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.098064 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.108924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.109301 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4jdt\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-kube-api-access-f4jdt\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.109434 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.109532 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-config-data\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.109629 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.109969 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.111115 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.111377 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.111667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.111832 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.111979 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.112086 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.119287 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.217799 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.218865 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-config-data\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.218904 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.218932 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-server-conf\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.218950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbxqc\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-kube-api-access-wbxqc\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.218973 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219736 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219781 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219799 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219821 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-server-conf\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219867 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d030be7f-cfb2-4335-a2f8-7da8f7430b17-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219894 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219915 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219936 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219960 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219978 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bef00f62-0142-4a72-9849-12246c13d004-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.219996 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-config-data\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220055 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220087 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hhmz\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-kube-api-access-7hhmz\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220106 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220132 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220150 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220188 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220208 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d030be7f-cfb2-4335-a2f8-7da8f7430b17-pod-info\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4jdt\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-kube-api-access-f4jdt\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220272 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220290 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-config-data\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220306 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220328 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220346 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220364 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bef00f62-0142-4a72-9849-12246c13d004-pod-info\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220384 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.220776 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.221495 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-config-data\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.221591 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.222074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.222549 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.226779 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.226857 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.227275 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.227312 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4922a1e397587981e3404408ee9f34eec949eb29c3877b1b5adb0ae562e7c69f/globalmount\"" pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.227536 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.236688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" event={"ID":"0d545816-db6b-4953-99a2-716ed95439ac","Type":"ContainerStarted","Data":"d19600dbcdd675a8553b7c7817e8bca4a485c84e3a9d6b450b28326ef2ecc821"} Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.241913 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" event={"ID":"aa886d28-0827-4dae-84bf-a69f196a8933","Type":"ContainerStarted","Data":"f14cb111de740a1f15eefec48066f285d5995220b1361a81b7119f92a3d25a99"} Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.241935 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.244115 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4jdt\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-kube-api-access-f4jdt\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.266000 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") pod \"rabbitmq-server-0\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322149 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-server-conf\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322272 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d030be7f-cfb2-4335-a2f8-7da8f7430b17-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322300 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322321 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322339 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bef00f62-0142-4a72-9849-12246c13d004-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322354 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-config-data\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322389 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322411 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hhmz\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-kube-api-access-7hhmz\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322447 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322463 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322479 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d030be7f-cfb2-4335-a2f8-7da8f7430b17-pod-info\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322515 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bef00f62-0142-4a72-9849-12246c13d004-pod-info\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322531 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322550 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322574 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-config-data\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322598 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322619 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-server-conf\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbxqc\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-kube-api-access-wbxqc\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322652 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.322685 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.323088 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.324004 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.325276 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-server-conf\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.334248 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.334290 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-config-data\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.334510 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.335041 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.335433 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.336282 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-config-data\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.337920 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.337925 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.337942 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d05509e9401e2174edbf89441304d15eb0beaf5fefd4c53cc6bfeba788bc2b73/globalmount\"" pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.337953 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fab1f2299b47ba0d1659e05dfaf91881496baf7a2af79f5e175e50cd9ff7b4c7/globalmount\"" pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.338404 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d030be7f-cfb2-4335-a2f8-7da8f7430b17-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.339092 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-server-conf\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.339186 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bef00f62-0142-4a72-9849-12246c13d004-pod-info\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.341769 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.342378 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bef00f62-0142-4a72-9849-12246c13d004-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.345401 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d030be7f-cfb2-4335-a2f8-7da8f7430b17-pod-info\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.357326 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.359868 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.360095 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.360590 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hhmz\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-kube-api-access-7hhmz\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.364499 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.377954 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.381485 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.385794 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.385988 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gsssz" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.386029 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.386116 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.386230 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.386292 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.393671 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.399995 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.410002 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbxqc\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-kube-api-access-wbxqc\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.459431 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") pod \"rabbitmq-server-1\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.500693 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") pod \"rabbitmq-server-2\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538135 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538281 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538411 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538507 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538560 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rd5p\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-kube-api-access-9rd5p\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538628 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538666 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538715 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.538740 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.640506 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.640833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.640893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.640928 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.640952 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.641070 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.641098 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rd5p\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-kube-api-access-9rd5p\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.641149 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.641188 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.641231 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.641255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.644749 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.645050 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.647533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.647549 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.648295 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.649403 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.650801 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.654851 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.658206 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.658719 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.658747 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/86ab0050a515986cbb8c4600c50553d9499b965b1a548dd8be0d02f2d54017c7/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.670209 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rd5p\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-kube-api-access-9rd5p\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.719085 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.727566 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Jan 06 14:55:58 crc kubenswrapper[4744]: I0106 14:55:58.797927 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") pod \"rabbitmq-cell1-server-0\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.047256 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.190238 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.429575 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.439078 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.441317 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.443362 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-kl2jz" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.444451 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.445992 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.448578 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.448698 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.450457 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.577118 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b2b4c15b-189d-41b2-b121-fdc25beb18a4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.577513 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2b4c15b-189d-41b2-b121-fdc25beb18a4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.577698 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a7781b1a-fc7e-438b-a64c-bccde1e62f83\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7781b1a-fc7e-438b-a64c-bccde1e62f83\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.577768 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b4c15b-189d-41b2-b121-fdc25beb18a4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.577980 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b2b4c15b-189d-41b2-b121-fdc25beb18a4-kolla-config\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.578108 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2b4c15b-189d-41b2-b121-fdc25beb18a4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.578353 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b2b4c15b-189d-41b2-b121-fdc25beb18a4-config-data-default\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.578520 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65hgg\" (UniqueName: \"kubernetes.io/projected/b2b4c15b-189d-41b2-b121-fdc25beb18a4-kube-api-access-65hgg\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.589189 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Jan 06 14:55:59 crc kubenswrapper[4744]: W0106 14:55:59.666215 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbef00f62_0142_4a72_9849_12246c13d004.slice/crio-183a0a8f321be07ca3156bc0309d0d413397f55c64de37ac737fa0da11c0ac2b WatchSource:0}: Error finding container 183a0a8f321be07ca3156bc0309d0d413397f55c64de37ac737fa0da11c0ac2b: Status 404 returned error can't find the container with id 183a0a8f321be07ca3156bc0309d0d413397f55c64de37ac737fa0da11c0ac2b Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.680320 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b4c15b-189d-41b2-b121-fdc25beb18a4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.681624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b2b4c15b-189d-41b2-b121-fdc25beb18a4-kolla-config\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.681649 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2b4c15b-189d-41b2-b121-fdc25beb18a4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.681774 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b2b4c15b-189d-41b2-b121-fdc25beb18a4-config-data-default\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.681818 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65hgg\" (UniqueName: \"kubernetes.io/projected/b2b4c15b-189d-41b2-b121-fdc25beb18a4-kube-api-access-65hgg\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.681895 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b2b4c15b-189d-41b2-b121-fdc25beb18a4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.681947 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2b4c15b-189d-41b2-b121-fdc25beb18a4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.681968 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a7781b1a-fc7e-438b-a64c-bccde1e62f83\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7781b1a-fc7e-438b-a64c-bccde1e62f83\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.682932 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b2b4c15b-189d-41b2-b121-fdc25beb18a4-config-data-default\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.684531 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b2b4c15b-189d-41b2-b121-fdc25beb18a4-kolla-config\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.684758 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b2b4c15b-189d-41b2-b121-fdc25beb18a4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.686840 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.686889 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a7781b1a-fc7e-438b-a64c-bccde1e62f83\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7781b1a-fc7e-438b-a64c-bccde1e62f83\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/92e3eb411e75ac413cec1fb6eabc93162984e61f71cceb3e1e1cb28dc3530141/globalmount\"" pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.688171 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2b4c15b-189d-41b2-b121-fdc25beb18a4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.689447 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2b4c15b-189d-41b2-b121-fdc25beb18a4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.690555 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b2b4c15b-189d-41b2-b121-fdc25beb18a4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.700895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65hgg\" (UniqueName: \"kubernetes.io/projected/b2b4c15b-189d-41b2-b121-fdc25beb18a4-kube-api-access-65hgg\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.773240 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a7781b1a-fc7e-438b-a64c-bccde1e62f83\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a7781b1a-fc7e-438b-a64c-bccde1e62f83\") pod \"openstack-galera-0\" (UID: \"b2b4c15b-189d-41b2-b121-fdc25beb18a4\") " pod="openstack/openstack-galera-0" Jan 06 14:55:59 crc kubenswrapper[4744]: I0106 14:55:59.793803 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.233420 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.290515 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"d030be7f-cfb2-4335-a2f8-7da8f7430b17","Type":"ContainerStarted","Data":"5496c5c0eb0ceb1a920069e3013be3ed96f77de0f5bd779d2316d842b284844c"} Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.292931 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff","Type":"ContainerStarted","Data":"43318afe9772285af4911e794bb68f4d82640a7d7352ed2508bcc0eafb7237a1"} Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.294912 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"bef00f62-0142-4a72-9849-12246c13d004","Type":"ContainerStarted","Data":"183a0a8f321be07ca3156bc0309d0d413397f55c64de37ac737fa0da11c0ac2b"} Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.297412 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4","Type":"ContainerStarted","Data":"d48c7f4b6fbbd43e4b4c36ab9bc02c0d6374fbf65fd6006c796b98ea5659dc90"} Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.372009 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 06 14:56:00 crc kubenswrapper[4744]: W0106 14:56:00.394017 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2b4c15b_189d_41b2_b121_fdc25beb18a4.slice/crio-05b3d6c4d00df5210359d8bfb730e192847791a659473806bd4bc22a77ffac89 WatchSource:0}: Error finding container 05b3d6c4d00df5210359d8bfb730e192847791a659473806bd4bc22a77ffac89: Status 404 returned error can't find the container with id 05b3d6c4d00df5210359d8bfb730e192847791a659473806bd4bc22a77ffac89 Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.627277 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.637006 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.640223 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-d7mtb" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.640516 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.640682 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.641452 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.644342 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.719868 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dw24\" (UniqueName: \"kubernetes.io/projected/bc13577b-cb8c-4e94-9109-34d00a75d148-kube-api-access-6dw24\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.719955 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bc13577b-cb8c-4e94-9109-34d00a75d148-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.719994 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bc13577b-cb8c-4e94-9109-34d00a75d148-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.720028 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bc13577b-cb8c-4e94-9109-34d00a75d148-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.720048 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc13577b-cb8c-4e94-9109-34d00a75d148-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.720136 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc13577b-cb8c-4e94-9109-34d00a75d148-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.720357 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc13577b-cb8c-4e94-9109-34d00a75d148-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.720448 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-69b2a3ef-6c47-445e-9c3a-74a9af609027\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69b2a3ef-6c47-445e-9c3a-74a9af609027\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.827335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bc13577b-cb8c-4e94-9109-34d00a75d148-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.827425 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bc13577b-cb8c-4e94-9109-34d00a75d148-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.827458 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bc13577b-cb8c-4e94-9109-34d00a75d148-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.827477 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc13577b-cb8c-4e94-9109-34d00a75d148-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.827500 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc13577b-cb8c-4e94-9109-34d00a75d148-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.827680 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc13577b-cb8c-4e94-9109-34d00a75d148-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.827707 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-69b2a3ef-6c47-445e-9c3a-74a9af609027\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69b2a3ef-6c47-445e-9c3a-74a9af609027\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.827829 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dw24\" (UniqueName: \"kubernetes.io/projected/bc13577b-cb8c-4e94-9109-34d00a75d148-kube-api-access-6dw24\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.831676 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/bc13577b-cb8c-4e94-9109-34d00a75d148-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.831924 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/bc13577b-cb8c-4e94-9109-34d00a75d148-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.832589 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/bc13577b-cb8c-4e94-9109-34d00a75d148-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.833920 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc13577b-cb8c-4e94-9109-34d00a75d148-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.838839 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc13577b-cb8c-4e94-9109-34d00a75d148-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.840458 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.840502 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-69b2a3ef-6c47-445e-9c3a-74a9af609027\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69b2a3ef-6c47-445e-9c3a-74a9af609027\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7b0cc7f16c0f9043fdb2f965941dfab3e1b3f7555d8bb33445f89e7d96118613/globalmount\"" pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.852891 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/bc13577b-cb8c-4e94-9109-34d00a75d148-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.889961 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dw24\" (UniqueName: \"kubernetes.io/projected/bc13577b-cb8c-4e94-9109-34d00a75d148-kube-api-access-6dw24\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.900425 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.902024 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.908493 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.908877 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-5h2js" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.909122 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Jan 06 14:56:00 crc kubenswrapper[4744]: I0106 14:56:00.910783 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.051265 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f28b7936-8bd1-41de-adba-74506e7df59e-config-data\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.051337 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f28b7936-8bd1-41de-adba-74506e7df59e-kolla-config\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.051374 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f28b7936-8bd1-41de-adba-74506e7df59e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.051551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f28b7936-8bd1-41de-adba-74506e7df59e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.051595 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbk9w\" (UniqueName: \"kubernetes.io/projected/f28b7936-8bd1-41de-adba-74506e7df59e-kube-api-access-mbk9w\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.058305 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-69b2a3ef-6c47-445e-9c3a-74a9af609027\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-69b2a3ef-6c47-445e-9c3a-74a9af609027\") pod \"openstack-cell1-galera-0\" (UID: \"bc13577b-cb8c-4e94-9109-34d00a75d148\") " pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.156169 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f28b7936-8bd1-41de-adba-74506e7df59e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.156229 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbk9w\" (UniqueName: \"kubernetes.io/projected/f28b7936-8bd1-41de-adba-74506e7df59e-kube-api-access-mbk9w\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.156320 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f28b7936-8bd1-41de-adba-74506e7df59e-config-data\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.156339 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f28b7936-8bd1-41de-adba-74506e7df59e-kolla-config\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.156362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f28b7936-8bd1-41de-adba-74506e7df59e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.163091 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f28b7936-8bd1-41de-adba-74506e7df59e-config-data\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.163698 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f28b7936-8bd1-41de-adba-74506e7df59e-kolla-config\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.165832 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f28b7936-8bd1-41de-adba-74506e7df59e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.176443 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f28b7936-8bd1-41de-adba-74506e7df59e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.185579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbk9w\" (UniqueName: \"kubernetes.io/projected/f28b7936-8bd1-41de-adba-74506e7df59e-kube-api-access-mbk9w\") pod \"memcached-0\" (UID: \"f28b7936-8bd1-41de-adba-74506e7df59e\") " pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.281681 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.295416 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:01 crc kubenswrapper[4744]: I0106 14:56:01.396322 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b2b4c15b-189d-41b2-b121-fdc25beb18a4","Type":"ContainerStarted","Data":"05b3d6c4d00df5210359d8bfb730e192847791a659473806bd4bc22a77ffac89"} Jan 06 14:56:03 crc kubenswrapper[4744]: I0106 14:56:03.036234 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 06 14:56:03 crc kubenswrapper[4744]: I0106 14:56:03.039175 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 06 14:56:03 crc kubenswrapper[4744]: I0106 14:56:03.044182 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-vxw9c" Jan 06 14:56:03 crc kubenswrapper[4744]: I0106 14:56:03.063640 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 06 14:56:03 crc kubenswrapper[4744]: I0106 14:56:03.144631 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnljm\" (UniqueName: \"kubernetes.io/projected/d0377d1a-719f-4733-bbd6-9cd135c1e764-kube-api-access-xnljm\") pod \"kube-state-metrics-0\" (UID: \"d0377d1a-719f-4733-bbd6-9cd135c1e764\") " pod="openstack/kube-state-metrics-0" Jan 06 14:56:03 crc kubenswrapper[4744]: I0106 14:56:03.247916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnljm\" (UniqueName: \"kubernetes.io/projected/d0377d1a-719f-4733-bbd6-9cd135c1e764-kube-api-access-xnljm\") pod \"kube-state-metrics-0\" (UID: \"d0377d1a-719f-4733-bbd6-9cd135c1e764\") " pod="openstack/kube-state-metrics-0" Jan 06 14:56:03 crc kubenswrapper[4744]: I0106 14:56:03.298303 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnljm\" (UniqueName: \"kubernetes.io/projected/d0377d1a-719f-4733-bbd6-9cd135c1e764-kube-api-access-xnljm\") pod \"kube-state-metrics-0\" (UID: \"d0377d1a-719f-4733-bbd6-9cd135c1e764\") " pod="openstack/kube-state-metrics-0" Jan 06 14:56:03 crc kubenswrapper[4744]: I0106 14:56:03.421701 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.165220 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt"] Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.166954 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.176855 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-vc2pw" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.177648 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.221555 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt"] Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.296129 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9l87\" (UniqueName: \"kubernetes.io/projected/b3368f37-f50b-431f-9900-c17002a24788-kube-api-access-x9l87\") pod \"observability-ui-dashboards-66cbf594b5-mbspt\" (UID: \"b3368f37-f50b-431f-9900-c17002a24788\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.296235 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3368f37-f50b-431f-9900-c17002a24788-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-mbspt\" (UID: \"b3368f37-f50b-431f-9900-c17002a24788\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.399408 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3368f37-f50b-431f-9900-c17002a24788-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-mbspt\" (UID: \"b3368f37-f50b-431f-9900-c17002a24788\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.399544 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9l87\" (UniqueName: \"kubernetes.io/projected/b3368f37-f50b-431f-9900-c17002a24788-kube-api-access-x9l87\") pod \"observability-ui-dashboards-66cbf594b5-mbspt\" (UID: \"b3368f37-f50b-431f-9900-c17002a24788\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.429368 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3368f37-f50b-431f-9900-c17002a24788-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-mbspt\" (UID: \"b3368f37-f50b-431f-9900-c17002a24788\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.446031 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.446791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9l87\" (UniqueName: \"kubernetes.io/projected/b3368f37-f50b-431f-9900-c17002a24788-kube-api-access-x9l87\") pod \"observability-ui-dashboards-66cbf594b5-mbspt\" (UID: \"b3368f37-f50b-431f-9900-c17002a24788\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.449548 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.454979 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-f8cnq" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.455280 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.455480 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.455622 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.455736 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.456017 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.456141 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.478684 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.489294 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.507767 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6c8ff568c-ksn2s"] Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.509057 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.530813 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.546311 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6c8ff568c-ksn2s"] Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605491 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-config\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605553 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605578 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605609 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-54250398-c52e-4c6a-9797-9a30efdd8220\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605631 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605719 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/702af028-b407-4750-a0bc-40fc8f144163-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605781 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zqt6\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-kube-api-access-8zqt6\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605844 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.605948 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707262 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-console-config\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qzqn\" (UniqueName: \"kubernetes.io/projected/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-kube-api-access-4qzqn\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707556 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-console-oauth-config\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707580 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-oauth-serving-cert\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707626 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zqt6\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-kube-api-access-8zqt6\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707657 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707680 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-service-ca\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707712 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707742 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707759 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-config\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707786 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-trusted-ca-bundle\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707829 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707846 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-console-serving-cert\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707871 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-54250398-c52e-4c6a-9797-9a30efdd8220\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.707952 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/702af028-b407-4750-a0bc-40fc8f144163-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.709085 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.709098 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.710469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.722009 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.723192 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.723230 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-54250398-c52e-4c6a-9797-9a30efdd8220\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7cd4730aba1d95fd82dde750f7c410cd74c9baee4ca03dd245d4d466c88db159/globalmount\"" pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.723534 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/702af028-b407-4750-a0bc-40fc8f144163-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.724244 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.727062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.736224 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-config\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.740317 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zqt6\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-kube-api-access-8zqt6\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.798404 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-54250398-c52e-4c6a-9797-9a30efdd8220\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220\") pod \"prometheus-metric-storage-0\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.809767 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-console-serving-cert\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.809957 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-console-config\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.809982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qzqn\" (UniqueName: \"kubernetes.io/projected/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-kube-api-access-4qzqn\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.810001 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-console-oauth-config\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.810044 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-oauth-serving-cert\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.810131 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-service-ca\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.810474 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-trusted-ca-bundle\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.811593 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-trusted-ca-bundle\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.812135 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-service-ca\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.812738 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-oauth-serving-cert\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.813070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-console-config\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.813193 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-console-oauth-config\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.815520 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-console-serving-cert\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.819401 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.833447 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qzqn\" (UniqueName: \"kubernetes.io/projected/f4dc35a3-9308-43dd-b1a5-4a13e6205b5f-kube-api-access-4qzqn\") pod \"console-6c8ff568c-ksn2s\" (UID: \"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f\") " pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:04 crc kubenswrapper[4744]: I0106 14:56:04.834848 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.128094 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.132187 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.134255 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.138475 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.138901 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-897b9" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.140567 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.140778 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.158575 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.274509 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.274568 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.274641 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.274661 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.274846 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr2tl\" (UniqueName: \"kubernetes.io/projected/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-kube-api-access-mr2tl\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.275044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f9d644ed-f595-4612-a0d0-a9993391a635\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f9d644ed-f595-4612-a0d0-a9993391a635\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.275076 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.275168 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-config\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.378938 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-f9d644ed-f595-4612-a0d0-a9993391a635\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f9d644ed-f595-4612-a0d0-a9993391a635\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.379020 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.381327 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.384926 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-config\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.385131 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.386668 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-config\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.387538 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.387763 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.387805 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.387954 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr2tl\" (UniqueName: \"kubernetes.io/projected/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-kube-api-access-mr2tl\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.390129 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.391023 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.391050 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-f9d644ed-f595-4612-a0d0-a9993391a635\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f9d644ed-f595-4612-a0d0-a9993391a635\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2439ddcde74430c86c51a770e0d2d09d129b09789ff6d19eeec2e1aab97c4eeb/globalmount\"" pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.395719 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.407885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.410102 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr2tl\" (UniqueName: \"kubernetes.io/projected/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-kube-api-access-mr2tl\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.432127 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab7e8638-3ee7-472b-a6de-a15b321e5fd5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.443389 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-r4r76"] Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.445041 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.455207 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.455467 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-zxh29" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.455743 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.465494 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-f9d644ed-f595-4612-a0d0-a9993391a635\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f9d644ed-f595-4612-a0d0-a9993391a635\") pod \"ovsdbserver-nb-0\" (UID: \"ab7e8638-3ee7-472b-a6de-a15b321e5fd5\") " pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.478229 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-qcph6"] Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.487376 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.495770 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-r4r76"] Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.522233 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-qcph6"] Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.594396 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78600e4f-de4a-4079-b849-b7b2319685b0-scripts\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.594446 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-etc-ovs\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.594476 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7f86f187-37db-488e-b6f6-22a5becacecd-var-log-ovn\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.594633 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f86f187-37db-488e-b6f6-22a5becacecd-scripts\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.594754 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f86f187-37db-488e-b6f6-22a5becacecd-var-run\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.594789 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sddk\" (UniqueName: \"kubernetes.io/projected/78600e4f-de4a-4079-b849-b7b2319685b0-kube-api-access-7sddk\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.594846 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc4nv\" (UniqueName: \"kubernetes.io/projected/7f86f187-37db-488e-b6f6-22a5becacecd-kube-api-access-cc4nv\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.595029 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-var-lib\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.595112 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f86f187-37db-488e-b6f6-22a5becacecd-combined-ca-bundle\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.595136 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-var-run\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.595239 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-var-log\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.595415 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f86f187-37db-488e-b6f6-22a5becacecd-ovn-controller-tls-certs\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.595538 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f86f187-37db-488e-b6f6-22a5becacecd-var-run-ovn\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.697716 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f86f187-37db-488e-b6f6-22a5becacecd-ovn-controller-tls-certs\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.697792 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f86f187-37db-488e-b6f6-22a5becacecd-var-run-ovn\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.697820 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78600e4f-de4a-4079-b849-b7b2319685b0-scripts\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.697841 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-etc-ovs\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.697860 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7f86f187-37db-488e-b6f6-22a5becacecd-var-log-ovn\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.697895 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f86f187-37db-488e-b6f6-22a5becacecd-scripts\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.697927 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f86f187-37db-488e-b6f6-22a5becacecd-var-run\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.697955 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sddk\" (UniqueName: \"kubernetes.io/projected/78600e4f-de4a-4079-b849-b7b2319685b0-kube-api-access-7sddk\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.697989 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc4nv\" (UniqueName: \"kubernetes.io/projected/7f86f187-37db-488e-b6f6-22a5becacecd-kube-api-access-cc4nv\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.698027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-var-lib\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.698056 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f86f187-37db-488e-b6f6-22a5becacecd-combined-ca-bundle\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.698072 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-var-run\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.698105 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-var-log\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.698352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7f86f187-37db-488e-b6f6-22a5becacecd-var-run-ovn\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.698409 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-var-log\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.698612 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-etc-ovs\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.698703 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7f86f187-37db-488e-b6f6-22a5becacecd-var-log-ovn\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.698775 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-var-lib\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.699444 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/78600e4f-de4a-4079-b849-b7b2319685b0-var-run\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.699694 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7f86f187-37db-488e-b6f6-22a5becacecd-var-run\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.699915 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/78600e4f-de4a-4079-b849-b7b2319685b0-scripts\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.700605 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7f86f187-37db-488e-b6f6-22a5becacecd-scripts\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.702520 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f86f187-37db-488e-b6f6-22a5becacecd-ovn-controller-tls-certs\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.709977 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f86f187-37db-488e-b6f6-22a5becacecd-combined-ca-bundle\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.717099 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc4nv\" (UniqueName: \"kubernetes.io/projected/7f86f187-37db-488e-b6f6-22a5becacecd-kube-api-access-cc4nv\") pod \"ovn-controller-r4r76\" (UID: \"7f86f187-37db-488e-b6f6-22a5becacecd\") " pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.717785 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sddk\" (UniqueName: \"kubernetes.io/projected/78600e4f-de4a-4079-b849-b7b2319685b0-kube-api-access-7sddk\") pod \"ovn-controller-ovs-qcph6\" (UID: \"78600e4f-de4a-4079-b849-b7b2319685b0\") " pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.772093 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.840433 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r4r76" Jan 06 14:56:07 crc kubenswrapper[4744]: I0106 14:56:07.861074 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.182642 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.184570 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.187412 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.187622 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-5t7f4" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.188885 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.188934 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.222514 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.358106 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c79025-4c6b-4040-9f3d-4b43390b9c17-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.358199 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/13c79025-4c6b-4040-9f3d-4b43390b9c17-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.358235 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13c79025-4c6b-4040-9f3d-4b43390b9c17-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.358299 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8ce9c3bf-833d-42f8-9883-85d31b48f4ae\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8ce9c3bf-833d-42f8-9883-85d31b48f4ae\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.360468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c79025-4c6b-4040-9f3d-4b43390b9c17-config\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.360620 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/13c79025-4c6b-4040-9f3d-4b43390b9c17-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.360721 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/13c79025-4c6b-4040-9f3d-4b43390b9c17-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.360780 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzspn\" (UniqueName: \"kubernetes.io/projected/13c79025-4c6b-4040-9f3d-4b43390b9c17-kube-api-access-fzspn\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.465524 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c79025-4c6b-4040-9f3d-4b43390b9c17-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.465618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/13c79025-4c6b-4040-9f3d-4b43390b9c17-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.465661 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13c79025-4c6b-4040-9f3d-4b43390b9c17-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.465768 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8ce9c3bf-833d-42f8-9883-85d31b48f4ae\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8ce9c3bf-833d-42f8-9883-85d31b48f4ae\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.465913 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c79025-4c6b-4040-9f3d-4b43390b9c17-config\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.465944 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/13c79025-4c6b-4040-9f3d-4b43390b9c17-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.465990 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/13c79025-4c6b-4040-9f3d-4b43390b9c17-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.466029 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzspn\" (UniqueName: \"kubernetes.io/projected/13c79025-4c6b-4040-9f3d-4b43390b9c17-kube-api-access-fzspn\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.467682 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/13c79025-4c6b-4040-9f3d-4b43390b9c17-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.468063 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c79025-4c6b-4040-9f3d-4b43390b9c17-config\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.470790 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13c79025-4c6b-4040-9f3d-4b43390b9c17-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.473785 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/13c79025-4c6b-4040-9f3d-4b43390b9c17-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.481131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/13c79025-4c6b-4040-9f3d-4b43390b9c17-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.481289 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.481324 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8ce9c3bf-833d-42f8-9883-85d31b48f4ae\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8ce9c3bf-833d-42f8-9883-85d31b48f4ae\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/443a73db6c22e3164f6abdc2ce9a58f4cc8592710e94d4dfd98b2f38c0ba2cc1/globalmount\"" pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.481340 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13c79025-4c6b-4040-9f3d-4b43390b9c17-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.485605 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzspn\" (UniqueName: \"kubernetes.io/projected/13c79025-4c6b-4040-9f3d-4b43390b9c17-kube-api-access-fzspn\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.529923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8ce9c3bf-833d-42f8-9883-85d31b48f4ae\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8ce9c3bf-833d-42f8-9883-85d31b48f4ae\") pod \"ovsdbserver-sb-0\" (UID: \"13c79025-4c6b-4040-9f3d-4b43390b9c17\") " pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:10 crc kubenswrapper[4744]: I0106 14:56:10.816969 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:21 crc kubenswrapper[4744]: E0106 14:56:21.369925 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 06 14:56:21 crc kubenswrapper[4744]: E0106 14:56:21.370603 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c67bp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-tqpfk_openstack(a1ab7c26-7c69-43d2-a237-085cd33418c2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:56:21 crc kubenswrapper[4744]: E0106 14:56:21.371749 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" podUID="a1ab7c26-7c69-43d2-a237-085cd33418c2" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.297722 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.298297 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qjwgm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-tmvg4_openstack(6c7bc1b1-1be0-45d5-81e4-ca4906b814ee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.299720 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" podUID="6c7bc1b1-1be0-45d5-81e4-ca4906b814ee" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.337706 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.337977 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4tv7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-mlnxx_openstack(0d545816-db6b-4953-99a2-716ed95439ac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.340907 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" podUID="0d545816-db6b-4953-99a2-716ed95439ac" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.377627 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.378097 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4sp5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-cmsvx_openstack(aa886d28-0827-4dae-84bf-a69f196a8933): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.379933 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" podUID="aa886d28-0827-4dae-84bf-a69f196a8933" Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.489201 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.688897 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c67bp\" (UniqueName: \"kubernetes.io/projected/a1ab7c26-7c69-43d2-a237-085cd33418c2-kube-api-access-c67bp\") pod \"a1ab7c26-7c69-43d2-a237-085cd33418c2\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.689360 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-config\") pod \"a1ab7c26-7c69-43d2-a237-085cd33418c2\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.689387 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-dns-svc\") pod \"a1ab7c26-7c69-43d2-a237-085cd33418c2\" (UID: \"a1ab7c26-7c69-43d2-a237-085cd33418c2\") " Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.690552 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a1ab7c26-7c69-43d2-a237-085cd33418c2" (UID: "a1ab7c26-7c69-43d2-a237-085cd33418c2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.690966 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-config" (OuterVolumeSpecName: "config") pod "a1ab7c26-7c69-43d2-a237-085cd33418c2" (UID: "a1ab7c26-7c69-43d2-a237-085cd33418c2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.700054 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ab7c26-7c69-43d2-a237-085cd33418c2-kube-api-access-c67bp" (OuterVolumeSpecName: "kube-api-access-c67bp") pod "a1ab7c26-7c69-43d2-a237-085cd33418c2" (UID: "a1ab7c26-7c69-43d2-a237-085cd33418c2"). InnerVolumeSpecName "kube-api-access-c67bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.731371 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.733208 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tqpfk" event={"ID":"a1ab7c26-7c69-43d2-a237-085cd33418c2","Type":"ContainerDied","Data":"aab4c6a3bfafba9980dfac788f7537d996199f0ab510eb0eb31d543f46fb0a57"} Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.733318 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" podUID="0d545816-db6b-4953-99a2-716ed95439ac" Jan 06 14:56:23 crc kubenswrapper[4744]: E0106 14:56:23.733389 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" podUID="aa886d28-0827-4dae-84bf-a69f196a8933" Jan 06 14:56:23 crc kubenswrapper[4744]: W0106 14:56:23.791251 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf28b7936_8bd1_41de_adba_74506e7df59e.slice/crio-2fe7969cb94d33250609f0cee13cb13f1b749edf88a0d4ea424d3492e1e0670d WatchSource:0}: Error finding container 2fe7969cb94d33250609f0cee13cb13f1b749edf88a0d4ea424d3492e1e0670d: Status 404 returned error can't find the container with id 2fe7969cb94d33250609f0cee13cb13f1b749edf88a0d4ea424d3492e1e0670d Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.794413 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c67bp\" (UniqueName: \"kubernetes.io/projected/a1ab7c26-7c69-43d2-a237-085cd33418c2-kube-api-access-c67bp\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.794455 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.794468 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ab7c26-7c69-43d2-a237-085cd33418c2-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.809581 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.812236 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.824954 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6c8ff568c-ksn2s"] Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.880918 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tqpfk"] Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.893713 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tqpfk"] Jan 06 14:56:23 crc kubenswrapper[4744]: I0106 14:56:23.994340 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.597309 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt"] Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.605596 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-r4r76"] Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.613053 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 06 14:56:24 crc kubenswrapper[4744]: W0106 14:56:24.629137 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3368f37_f50b_431f_9900_c17002a24788.slice/crio-776c04d19a2465f20d2a8d03b8e6502036aeffca577f6916e0cdb4eab6bb50bc WatchSource:0}: Error finding container 776c04d19a2465f20d2a8d03b8e6502036aeffca577f6916e0cdb4eab6bb50bc: Status 404 returned error can't find the container with id 776c04d19a2465f20d2a8d03b8e6502036aeffca577f6916e0cdb4eab6bb50bc Jan 06 14:56:24 crc kubenswrapper[4744]: W0106 14:56:24.636582 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0377d1a_719f_4733_bbd6_9cd135c1e764.slice/crio-fae50407dc8919f66deac594799250e2934ff52c111ad5467423db5677615ea6 WatchSource:0}: Error finding container fae50407dc8919f66deac594799250e2934ff52c111ad5467423db5677615ea6: Status 404 returned error can't find the container with id fae50407dc8919f66deac594799250e2934ff52c111ad5467423db5677615ea6 Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.753341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r4r76" event={"ID":"7f86f187-37db-488e-b6f6-22a5becacecd","Type":"ContainerStarted","Data":"993ab1e06ac19b6133b7d972da8459a001e18d5b42368d9e006269cc0a1d6820"} Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.765573 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerStarted","Data":"d26a51cabb330d65bcf8efc36f345d2d133e669329ec7cd912284dd9e92b8e93"} Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.766700 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" event={"ID":"b3368f37-f50b-431f-9900-c17002a24788","Type":"ContainerStarted","Data":"776c04d19a2465f20d2a8d03b8e6502036aeffca577f6916e0cdb4eab6bb50bc"} Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.780102 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6c8ff568c-ksn2s" event={"ID":"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f","Type":"ContainerStarted","Data":"ef47e3478a5957dbfca28ac416484b0d25a574f0feb9d0269173fd339f295e26"} Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.780144 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6c8ff568c-ksn2s" event={"ID":"f4dc35a3-9308-43dd-b1a5-4a13e6205b5f","Type":"ContainerStarted","Data":"03d802a6340193f6896d60575d1bcff2850152147edb63e7f3a716211d411c00"} Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.786729 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.807770 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d0377d1a-719f-4733-bbd6-9cd135c1e764","Type":"ContainerStarted","Data":"fae50407dc8919f66deac594799250e2934ff52c111ad5467423db5677615ea6"} Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.810210 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b2b4c15b-189d-41b2-b121-fdc25beb18a4","Type":"ContainerStarted","Data":"4191e40b45170ba8468773792a6cda0f270562851b04fd47fcf38a567ee29e69"} Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.810872 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6c8ff568c-ksn2s" podStartSLOduration=20.810855726 podStartE2EDuration="20.810855726s" podCreationTimestamp="2026-01-06 14:56:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:56:24.807011915 +0000 UTC m=+1181.434478233" watchObservedRunningTime="2026-01-06 14:56:24.810855726 +0000 UTC m=+1181.438322044" Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.812578 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f28b7936-8bd1-41de-adba-74506e7df59e","Type":"ContainerStarted","Data":"2fe7969cb94d33250609f0cee13cb13f1b749edf88a0d4ea424d3492e1e0670d"} Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.815567 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" event={"ID":"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee","Type":"ContainerDied","Data":"138a5a73da26dbf2bbd8765e595917658a85f391fbc6df87f89f426ceecdf3e9"} Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.815595 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="138a5a73da26dbf2bbd8765e595917658a85f391fbc6df87f89f426ceecdf3e9" Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.835862 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.835915 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.843985 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:24 crc kubenswrapper[4744]: I0106 14:56:24.949584 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-qcph6"] Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.104109 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.227836 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjwgm\" (UniqueName: \"kubernetes.io/projected/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-kube-api-access-qjwgm\") pod \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\" (UID: \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\") " Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.228518 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-config\") pod \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\" (UID: \"6c7bc1b1-1be0-45d5-81e4-ca4906b814ee\") " Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.229143 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-config" (OuterVolumeSpecName: "config") pod "6c7bc1b1-1be0-45d5-81e4-ca4906b814ee" (UID: "6c7bc1b1-1be0-45d5-81e4-ca4906b814ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.245491 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-kube-api-access-qjwgm" (OuterVolumeSpecName: "kube-api-access-qjwgm") pod "6c7bc1b1-1be0-45d5-81e4-ca4906b814ee" (UID: "6c7bc1b1-1be0-45d5-81e4-ca4906b814ee"). InnerVolumeSpecName "kube-api-access-qjwgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.331502 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.331532 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjwgm\" (UniqueName: \"kubernetes.io/projected/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee-kube-api-access-qjwgm\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.545368 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.724385 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1ab7c26-7c69-43d2-a237-085cd33418c2" path="/var/lib/kubelet/pods/a1ab7c26-7c69-43d2-a237-085cd33418c2/volumes" Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.794741 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.830105 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4","Type":"ContainerStarted","Data":"23858f13f4cc753d4995c5de460b682039086270f2029befd5f5afb9b8976ea7"} Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.834490 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bc13577b-cb8c-4e94-9109-34d00a75d148","Type":"ContainerStarted","Data":"fe88fe781c24a877cfd24a6928caa6143fa4603d5a48684606e095712cb54b19"} Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.834518 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bc13577b-cb8c-4e94-9109-34d00a75d148","Type":"ContainerStarted","Data":"cb64a55665f9cca6e87561c505f2bf2de7f8b88d85a06470d611dee4e568efaf"} Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.836864 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"d030be7f-cfb2-4335-a2f8-7da8f7430b17","Type":"ContainerStarted","Data":"7bdc826178224ec8f48dc20281097da9db5ad81ce52b8030850b40e45948d25b"} Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.841582 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff","Type":"ContainerStarted","Data":"b568e5b0f4450110f1563bd5f836d3c37170139a7c49b77607d38119a536ea25"} Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.844051 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"bef00f62-0142-4a72-9849-12246c13d004","Type":"ContainerStarted","Data":"738a503b4e10e99e3abe518b1919d9ce5bac03ecfcbea228a9694a955f0cf80d"} Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.845599 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qcph6" event={"ID":"78600e4f-de4a-4079-b849-b7b2319685b0","Type":"ContainerStarted","Data":"505c354025b6b4dd51b1bc83fb0b8380b27721dcf68d982f74c38c8ead809f24"} Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.845655 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tmvg4" Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.862310 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6c8ff568c-ksn2s" Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.902217 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tmvg4"] Jan 06 14:56:25 crc kubenswrapper[4744]: I0106 14:56:25.908706 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tmvg4"] Jan 06 14:56:26 crc kubenswrapper[4744]: I0106 14:56:26.072953 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-857bd77bfc-rvs27"] Jan 06 14:56:26 crc kubenswrapper[4744]: I0106 14:56:26.859812 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"13c79025-4c6b-4040-9f3d-4b43390b9c17","Type":"ContainerStarted","Data":"097dcd5162146303486abe4a8c9ed068540dc975cc1a37356bd58a1f319ec145"} Jan 06 14:56:26 crc kubenswrapper[4744]: I0106 14:56:26.865355 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ab7e8638-3ee7-472b-a6de-a15b321e5fd5","Type":"ContainerStarted","Data":"95450522705ede3fc1dff9ae8105b0ad246485a1f459fe119545df22937a0a2c"} Jan 06 14:56:27 crc kubenswrapper[4744]: I0106 14:56:27.721205 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c7bc1b1-1be0-45d5-81e4-ca4906b814ee" path="/var/lib/kubelet/pods/6c7bc1b1-1be0-45d5-81e4-ca4906b814ee/volumes" Jan 06 14:56:27 crc kubenswrapper[4744]: I0106 14:56:27.874890 4744 generic.go:334] "Generic (PLEG): container finished" podID="b2b4c15b-189d-41b2-b121-fdc25beb18a4" containerID="4191e40b45170ba8468773792a6cda0f270562851b04fd47fcf38a567ee29e69" exitCode=0 Jan 06 14:56:27 crc kubenswrapper[4744]: I0106 14:56:27.874992 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b2b4c15b-189d-41b2-b121-fdc25beb18a4","Type":"ContainerDied","Data":"4191e40b45170ba8468773792a6cda0f270562851b04fd47fcf38a567ee29e69"} Jan 06 14:56:29 crc kubenswrapper[4744]: I0106 14:56:29.902589 4744 generic.go:334] "Generic (PLEG): container finished" podID="bc13577b-cb8c-4e94-9109-34d00a75d148" containerID="fe88fe781c24a877cfd24a6928caa6143fa4603d5a48684606e095712cb54b19" exitCode=0 Jan 06 14:56:29 crc kubenswrapper[4744]: I0106 14:56:29.902675 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bc13577b-cb8c-4e94-9109-34d00a75d148","Type":"ContainerDied","Data":"fe88fe781c24a877cfd24a6928caa6143fa4603d5a48684606e095712cb54b19"} Jan 06 14:56:36 crc kubenswrapper[4744]: E0106 14:56:36.825282 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Jan 06 14:56:36 crc kubenswrapper[4744]: E0106 14:56:36.825919 4744 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0" Jan 06 14:56:36 crc kubenswrapper[4744]: E0106 14:56:36.826101 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-state-metrics,Image:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,Command:[],Args:[--resources=pods --namespaces=openstack],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},ContainerPort{Name:telemetry,HostPort:0,ContainerPort:8081,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xnljm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{0 8080 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod kube-state-metrics-0_openstack(d0377d1a-719f-4733-bbd6-9cd135c1e764): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 06 14:56:36 crc kubenswrapper[4744]: E0106 14:56:36.827329 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/kube-state-metrics-0" podUID="d0377d1a-719f-4733-bbd6-9cd135c1e764" Jan 06 14:56:36 crc kubenswrapper[4744]: I0106 14:56:36.981465 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f28b7936-8bd1-41de-adba-74506e7df59e","Type":"ContainerStarted","Data":"46558a60000060c912380df35d08042e2820880f21c5be9d32464d4915946a6d"} Jan 06 14:56:36 crc kubenswrapper[4744]: E0106 14:56:36.983032 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0\\\"\"" pod="openstack/kube-state-metrics-0" podUID="d0377d1a-719f-4733-bbd6-9cd135c1e764" Jan 06 14:56:37 crc kubenswrapper[4744]: I0106 14:56:37.993593 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"bc13577b-cb8c-4e94-9109-34d00a75d148","Type":"ContainerStarted","Data":"528c8c9335141a2b5ff15095b0ad0c140dbf50ced8f6fafb094389d2abb6fcd3"} Jan 06 14:56:37 crc kubenswrapper[4744]: I0106 14:56:37.997756 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ab7e8638-3ee7-472b-a6de-a15b321e5fd5","Type":"ContainerStarted","Data":"fafe91981ffb59681589437bc45c8cc42e21b0a589113c3f72ed76d67ad22214"} Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.000017 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"b2b4c15b-189d-41b2-b121-fdc25beb18a4","Type":"ContainerStarted","Data":"7f148ad471ec9f331ad02e531a5de9be6141a2b2aae541d0bc92f517f399bcde"} Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.002386 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qcph6" event={"ID":"78600e4f-de4a-4079-b849-b7b2319685b0","Type":"ContainerStarted","Data":"e2b673f698080ddb5febaf68814a1f6b6e357c8f85d2e8ad5ad0feb11ecd48db"} Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.011245 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"13c79025-4c6b-4040-9f3d-4b43390b9c17","Type":"ContainerStarted","Data":"16c6edd2e4a0164b9587920dbb3d1d4f575bda020a8d839ae196243a65d76574"} Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.014064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r4r76" event={"ID":"7f86f187-37db-488e-b6f6-22a5becacecd","Type":"ContainerStarted","Data":"bdd8650768a97c4cbddbd669af951ab7112352dcf1829e832c73a418f65c4422"} Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.022958 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" event={"ID":"b3368f37-f50b-431f-9900-c17002a24788","Type":"ContainerStarted","Data":"fde1e4edbe85a1d83d7f584deffd24e2bc1f8cc949d8d8239eadfd3dcdcdeb7f"} Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.023284 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.033796 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=39.033772221 podStartE2EDuration="39.033772221s" podCreationTimestamp="2026-01-06 14:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:56:38.02089285 +0000 UTC m=+1194.648359218" watchObservedRunningTime="2026-01-06 14:56:38.033772221 +0000 UTC m=+1194.661238559" Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.077511 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=16.902929566 podStartE2EDuration="40.077490677s" podCreationTimestamp="2026-01-06 14:55:58 +0000 UTC" firstStartedPulling="2026-01-06 14:56:00.405200789 +0000 UTC m=+1157.032667117" lastFinishedPulling="2026-01-06 14:56:23.57976191 +0000 UTC m=+1180.207228228" observedRunningTime="2026-01-06 14:56:38.051139 +0000 UTC m=+1194.678605328" watchObservedRunningTime="2026-01-06 14:56:38.077490677 +0000 UTC m=+1194.704957005" Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.083246 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=31.31326782 podStartE2EDuration="38.083235139s" podCreationTimestamp="2026-01-06 14:56:00 +0000 UTC" firstStartedPulling="2026-01-06 14:56:23.811872518 +0000 UTC m=+1180.439338836" lastFinishedPulling="2026-01-06 14:56:30.581839837 +0000 UTC m=+1187.209306155" observedRunningTime="2026-01-06 14:56:38.071814167 +0000 UTC m=+1194.699280495" watchObservedRunningTime="2026-01-06 14:56:38.083235139 +0000 UTC m=+1194.710701457" Jan 06 14:56:38 crc kubenswrapper[4744]: I0106 14:56:38.093618 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-mbspt" podStartSLOduration=25.439485089 podStartE2EDuration="34.093594773s" podCreationTimestamp="2026-01-06 14:56:04 +0000 UTC" firstStartedPulling="2026-01-06 14:56:24.638305163 +0000 UTC m=+1181.265771481" lastFinishedPulling="2026-01-06 14:56:33.292414807 +0000 UTC m=+1189.919881165" observedRunningTime="2026-01-06 14:56:38.086038733 +0000 UTC m=+1194.713505061" watchObservedRunningTime="2026-01-06 14:56:38.093594773 +0000 UTC m=+1194.721061091" Jan 06 14:56:39 crc kubenswrapper[4744]: I0106 14:56:39.794272 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Jan 06 14:56:39 crc kubenswrapper[4744]: I0106 14:56:39.794685 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Jan 06 14:56:40 crc kubenswrapper[4744]: I0106 14:56:40.052069 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-r4r76" Jan 06 14:56:40 crc kubenswrapper[4744]: I0106 14:56:40.085277 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-r4r76" podStartSLOduration=24.409997328 podStartE2EDuration="33.085245781s" podCreationTimestamp="2026-01-06 14:56:07 +0000 UTC" firstStartedPulling="2026-01-06 14:56:24.617073762 +0000 UTC m=+1181.244540080" lastFinishedPulling="2026-01-06 14:56:33.292322225 +0000 UTC m=+1189.919788533" observedRunningTime="2026-01-06 14:56:40.073579343 +0000 UTC m=+1196.701045691" watchObservedRunningTime="2026-01-06 14:56:40.085245781 +0000 UTC m=+1196.712712119" Jan 06 14:56:41 crc kubenswrapper[4744]: I0106 14:56:41.283563 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Jan 06 14:56:41 crc kubenswrapper[4744]: I0106 14:56:41.296188 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:41 crc kubenswrapper[4744]: I0106 14:56:41.296269 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.477067 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mlnxx"] Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.523749 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-kmzv6"] Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.525503 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.553861 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-kmzv6"] Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.661881 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ltld\" (UniqueName: \"kubernetes.io/projected/61664168-97f1-4e19-920a-d584ea5074b4-kube-api-access-2ltld\") pod \"dnsmasq-dns-7cb5889db5-kmzv6\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.662019 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-config\") pod \"dnsmasq-dns-7cb5889db5-kmzv6\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.662131 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-kmzv6\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.764529 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ltld\" (UniqueName: \"kubernetes.io/projected/61664168-97f1-4e19-920a-d584ea5074b4-kube-api-access-2ltld\") pod \"dnsmasq-dns-7cb5889db5-kmzv6\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.764646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-config\") pod \"dnsmasq-dns-7cb5889db5-kmzv6\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.764773 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-kmzv6\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.766879 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-config\") pod \"dnsmasq-dns-7cb5889db5-kmzv6\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.768302 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-kmzv6\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.784689 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ltld\" (UniqueName: \"kubernetes.io/projected/61664168-97f1-4e19-920a-d584ea5074b4-kube-api-access-2ltld\") pod \"dnsmasq-dns-7cb5889db5-kmzv6\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:43 crc kubenswrapper[4744]: I0106 14:56:43.856707 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.361358 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-kmzv6"] Jan 06 14:56:44 crc kubenswrapper[4744]: W0106 14:56:44.369447 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61664168_97f1_4e19_920a_d584ea5074b4.slice/crio-d809eaf79b810e50bfb6383d5e296353530653c9a0150b286a6aaadb2b415ee2 WatchSource:0}: Error finding container d809eaf79b810e50bfb6383d5e296353530653c9a0150b286a6aaadb2b415ee2: Status 404 returned error can't find the container with id d809eaf79b810e50bfb6383d5e296353530653c9a0150b286a6aaadb2b415ee2 Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.580863 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.622916 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.625771 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.628454 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.628626 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.629131 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-mrh8f" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.630651 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.700254 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6whn\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-kube-api-access-p6whn\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.703002 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-cache\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.703215 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.703669 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2b705766-7322-4b39-9298-946097bb9074\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2b705766-7322-4b39-9298-946097bb9074\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.703708 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-lock\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.812827 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2b705766-7322-4b39-9298-946097bb9074\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2b705766-7322-4b39-9298-946097bb9074\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.814555 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-lock\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.814804 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6whn\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-kube-api-access-p6whn\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.814924 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-cache\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.815226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: E0106 14:56:44.815417 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 06 14:56:44 crc kubenswrapper[4744]: E0106 14:56:44.815437 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.815449 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-cache\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: E0106 14:56:44.815482 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift podName:c603cde3-b7b0-4851-99df-2abb0aa6b1c4 nodeName:}" failed. No retries permitted until 2026-01-06 14:56:45.315469219 +0000 UTC m=+1201.942935537 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift") pod "swift-storage-0" (UID: "c603cde3-b7b0-4851-99df-2abb0aa6b1c4") : configmap "swift-ring-files" not found Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.815739 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-lock\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.825729 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.825776 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2b705766-7322-4b39-9298-946097bb9074\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2b705766-7322-4b39-9298-946097bb9074\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/416bec1bcbd299f7ed7de050b1b9061a511fbc91b00045dd09ce00741b0b25e8/globalmount\"" pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.833143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6whn\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-kube-api-access-p6whn\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:44 crc kubenswrapper[4744]: I0106 14:56:44.879852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2b705766-7322-4b39-9298-946097bb9074\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2b705766-7322-4b39-9298-946097bb9074\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.119927 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-fq864"] Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.121828 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.127580 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" event={"ID":"61664168-97f1-4e19-920a-d584ea5074b4","Type":"ContainerStarted","Data":"d809eaf79b810e50bfb6383d5e296353530653c9a0150b286a6aaadb2b415ee2"} Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.127762 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.127761 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.128004 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.144665 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-fq864"] Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.231638 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-combined-ca-bundle\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.231698 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/82b9e690-d1d7-473d-99d1-20be1130dfd3-etc-swift\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.231741 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg6cf\" (UniqueName: \"kubernetes.io/projected/82b9e690-d1d7-473d-99d1-20be1130dfd3-kube-api-access-qg6cf\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.231907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-dispersionconf\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.231996 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-scripts\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.232051 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-swiftconf\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.232103 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-ring-data-devices\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.333840 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-dispersionconf\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.333920 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-scripts\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.333950 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-swiftconf\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.333995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-ring-data-devices\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.334041 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-combined-ca-bundle\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.334060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/82b9e690-d1d7-473d-99d1-20be1130dfd3-etc-swift\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.334085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg6cf\" (UniqueName: \"kubernetes.io/projected/82b9e690-d1d7-473d-99d1-20be1130dfd3-kube-api-access-qg6cf\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.334118 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.334995 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-scripts\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: E0106 14:56:45.335105 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 06 14:56:45 crc kubenswrapper[4744]: E0106 14:56:45.335129 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 06 14:56:45 crc kubenswrapper[4744]: E0106 14:56:45.335201 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift podName:c603cde3-b7b0-4851-99df-2abb0aa6b1c4 nodeName:}" failed. No retries permitted until 2026-01-06 14:56:46.335182343 +0000 UTC m=+1202.962648731 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift") pod "swift-storage-0" (UID: "c603cde3-b7b0-4851-99df-2abb0aa6b1c4") : configmap "swift-ring-files" not found Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.337778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-ring-data-devices\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.337939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/82b9e690-d1d7-473d-99d1-20be1130dfd3-etc-swift\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.346413 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-dispersionconf\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.348876 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-combined-ca-bundle\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.353314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg6cf\" (UniqueName: \"kubernetes.io/projected/82b9e690-d1d7-473d-99d1-20be1130dfd3-kube-api-access-qg6cf\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.354816 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-swiftconf\") pod \"swift-ring-rebalance-fq864\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:45 crc kubenswrapper[4744]: I0106 14:56:45.447859 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:56:46 crc kubenswrapper[4744]: I0106 14:56:46.353041 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:46 crc kubenswrapper[4744]: E0106 14:56:46.353447 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 06 14:56:46 crc kubenswrapper[4744]: E0106 14:56:46.353461 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 06 14:56:46 crc kubenswrapper[4744]: E0106 14:56:46.353505 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift podName:c603cde3-b7b0-4851-99df-2abb0aa6b1c4 nodeName:}" failed. No retries permitted until 2026-01-06 14:56:48.353492092 +0000 UTC m=+1204.980958400 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift") pod "swift-storage-0" (UID: "c603cde3-b7b0-4851-99df-2abb0aa6b1c4") : configmap "swift-ring-files" not found Jan 06 14:56:47 crc kubenswrapper[4744]: I0106 14:56:47.153105 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerStarted","Data":"bc5b62258060bf20a1d4f1eb6935143bfd990dbab407532cfdf16ae29ab74621"} Jan 06 14:56:47 crc kubenswrapper[4744]: I0106 14:56:47.157414 4744 generic.go:334] "Generic (PLEG): container finished" podID="78600e4f-de4a-4079-b849-b7b2319685b0" containerID="e2b673f698080ddb5febaf68814a1f6b6e357c8f85d2e8ad5ad0feb11ecd48db" exitCode=0 Jan 06 14:56:47 crc kubenswrapper[4744]: I0106 14:56:47.157462 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qcph6" event={"ID":"78600e4f-de4a-4079-b849-b7b2319685b0","Type":"ContainerDied","Data":"e2b673f698080ddb5febaf68814a1f6b6e357c8f85d2e8ad5ad0feb11ecd48db"} Jan 06 14:56:48 crc kubenswrapper[4744]: I0106 14:56:48.401230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:48 crc kubenswrapper[4744]: E0106 14:56:48.401418 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 06 14:56:48 crc kubenswrapper[4744]: E0106 14:56:48.401619 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 06 14:56:48 crc kubenswrapper[4744]: E0106 14:56:48.401680 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift podName:c603cde3-b7b0-4851-99df-2abb0aa6b1c4 nodeName:}" failed. No retries permitted until 2026-01-06 14:56:52.401662755 +0000 UTC m=+1209.029129073 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift") pod "swift-storage-0" (UID: "c603cde3-b7b0-4851-99df-2abb0aa6b1c4") : configmap "swift-ring-files" not found Jan 06 14:56:49 crc kubenswrapper[4744]: E0106 14:56:49.234386 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.22:45692->38.102.83.22:45165: write tcp 38.102.83.22:45692->38.102.83.22:45165: write: connection reset by peer Jan 06 14:56:50 crc kubenswrapper[4744]: I0106 14:56:50.012011 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:50 crc kubenswrapper[4744]: I0106 14:56:50.101048 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Jan 06 14:56:50 crc kubenswrapper[4744]: I0106 14:56:50.224732 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-fq864"] Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.152366 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-857bd77bfc-rvs27" podUID="4b070bab-3e2c-4579-8d78-92ad1f973e62" containerName="console" containerID="cri-o://0f1a7c93518967526038315d0f2e18176895b43b2cfeeb421285a669ffb57ef8" gracePeriod=15 Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.212256 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fq864" event={"ID":"82b9e690-d1d7-473d-99d1-20be1130dfd3","Type":"ContainerStarted","Data":"ebb9c082be7a14f3d4235dfc16de25ee19935cf56bba639b9b0efbc811a29b35"} Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.215601 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qcph6" event={"ID":"78600e4f-de4a-4079-b849-b7b2319685b0","Type":"ContainerStarted","Data":"0bdcf6a40268f174c9ae75d8be749ea3fbd7a06105122ccf6cd99505d6f4a403"} Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.215626 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-qcph6" event={"ID":"78600e4f-de4a-4079-b849-b7b2319685b0","Type":"ContainerStarted","Data":"b7b5ca65d964fa4d7c6a0f19d548dc2856db69753ae8e973d8101b83b52fd342"} Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.215854 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.215898 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.218797 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"13c79025-4c6b-4040-9f3d-4b43390b9c17","Type":"ContainerStarted","Data":"5af7684d8deba11c3d79dd31cf29f1d87c4d9da52ff36aee253accbfc476e5c3"} Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.220838 4744 generic.go:334] "Generic (PLEG): container finished" podID="0d545816-db6b-4953-99a2-716ed95439ac" containerID="2928d12e981531521cd7af0cbf7c185997a971cb2860ae7c53a29e35d7850dd2" exitCode=0 Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.220892 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" event={"ID":"0d545816-db6b-4953-99a2-716ed95439ac","Type":"ContainerDied","Data":"2928d12e981531521cd7af0cbf7c185997a971cb2860ae7c53a29e35d7850dd2"} Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.225215 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ab7e8638-3ee7-472b-a6de-a15b321e5fd5","Type":"ContainerStarted","Data":"bb1ab597213faf1a59961193b3bf2d5670e92746bd2e2c3c274d788f216dd910"} Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.231147 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d0377d1a-719f-4733-bbd6-9cd135c1e764","Type":"ContainerStarted","Data":"50186dd7a4c6f1dd389ce2ccd5ea4f266bf4859e95c81c53042661c37d6ca91e"} Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.231410 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.233807 4744 generic.go:334] "Generic (PLEG): container finished" podID="aa886d28-0827-4dae-84bf-a69f196a8933" containerID="d7d54a92a83860382f5a94d088a01b8d96650c1308e8c59070bdb14535f45d09" exitCode=0 Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.233878 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" event={"ID":"aa886d28-0827-4dae-84bf-a69f196a8933","Type":"ContainerDied","Data":"d7d54a92a83860382f5a94d088a01b8d96650c1308e8c59070bdb14535f45d09"} Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.235771 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-qcph6" podStartSLOduration=36.557492893 podStartE2EDuration="44.235751281s" podCreationTimestamp="2026-01-06 14:56:07 +0000 UTC" firstStartedPulling="2026-01-06 14:56:24.950488489 +0000 UTC m=+1181.577954807" lastFinishedPulling="2026-01-06 14:56:32.628746837 +0000 UTC m=+1189.256213195" observedRunningTime="2026-01-06 14:56:51.235441133 +0000 UTC m=+1207.862907491" watchObservedRunningTime="2026-01-06 14:56:51.235751281 +0000 UTC m=+1207.863217599" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.245792 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" event={"ID":"61664168-97f1-4e19-920a-d584ea5074b4","Type":"ContainerDied","Data":"814b24f3776dc8599ba6be2125e8fc7c87a8bafce5ef853eab7d926a7d0fa25f"} Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.245796 4744 generic.go:334] "Generic (PLEG): container finished" podID="61664168-97f1-4e19-920a-d584ea5074b4" containerID="814b24f3776dc8599ba6be2125e8fc7c87a8bafce5ef853eab7d926a7d0fa25f" exitCode=0 Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.316757 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=21.208330515 podStartE2EDuration="45.316731813s" podCreationTimestamp="2026-01-06 14:56:06 +0000 UTC" firstStartedPulling="2026-01-06 14:56:25.975385141 +0000 UTC m=+1182.602851459" lastFinishedPulling="2026-01-06 14:56:50.083786439 +0000 UTC m=+1206.711252757" observedRunningTime="2026-01-06 14:56:51.289860562 +0000 UTC m=+1207.917326890" watchObservedRunningTime="2026-01-06 14:56:51.316731813 +0000 UTC m=+1207.944198131" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.344888 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-nsnv2"] Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.346735 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.350745 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.361434 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nsnv2"] Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.390735 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=18.354469749 podStartE2EDuration="42.390716069s" podCreationTimestamp="2026-01-06 14:56:09 +0000 UTC" firstStartedPulling="2026-01-06 14:56:25.975386361 +0000 UTC m=+1182.602852679" lastFinishedPulling="2026-01-06 14:56:50.011632681 +0000 UTC m=+1206.639098999" observedRunningTime="2026-01-06 14:56:51.389517698 +0000 UTC m=+1208.016984016" watchObservedRunningTime="2026-01-06 14:56:51.390716069 +0000 UTC m=+1208.018182387" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.414175 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=22.977442395 podStartE2EDuration="48.414139389s" podCreationTimestamp="2026-01-06 14:56:03 +0000 UTC" firstStartedPulling="2026-01-06 14:56:24.653741041 +0000 UTC m=+1181.281207359" lastFinishedPulling="2026-01-06 14:56:50.090438035 +0000 UTC m=+1206.717904353" observedRunningTime="2026-01-06 14:56:51.412967418 +0000 UTC m=+1208.040433746" watchObservedRunningTime="2026-01-06 14:56:51.414139389 +0000 UTC m=+1208.041605707" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.491578 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4kgj\" (UniqueName: \"kubernetes.io/projected/7e6b0b81-7dd2-4030-827e-54bbc1a59635-kube-api-access-g4kgj\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.491633 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7e6b0b81-7dd2-4030-827e-54bbc1a59635-ovs-rundir\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.491735 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e6b0b81-7dd2-4030-827e-54bbc1a59635-config\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.491823 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e6b0b81-7dd2-4030-827e-54bbc1a59635-combined-ca-bundle\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.491847 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e6b0b81-7dd2-4030-827e-54bbc1a59635-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.491920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7e6b0b81-7dd2-4030-827e-54bbc1a59635-ovn-rundir\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.510349 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cmsvx"] Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.537059 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-4gtkg"] Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.538717 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.545778 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.596801 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-dns-svc\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.596881 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7e6b0b81-7dd2-4030-827e-54bbc1a59635-ovn-rundir\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.597004 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4kgj\" (UniqueName: \"kubernetes.io/projected/7e6b0b81-7dd2-4030-827e-54bbc1a59635-kube-api-access-g4kgj\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.597046 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7e6b0b81-7dd2-4030-827e-54bbc1a59635-ovs-rundir\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.597098 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.597293 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2zkm\" (UniqueName: \"kubernetes.io/projected/7f9b819c-e185-4a98-8f78-32951d751132-kube-api-access-m2zkm\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.597383 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e6b0b81-7dd2-4030-827e-54bbc1a59635-config\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.597568 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-config\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.597646 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e6b0b81-7dd2-4030-827e-54bbc1a59635-combined-ca-bundle\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.597717 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e6b0b81-7dd2-4030-827e-54bbc1a59635-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.599528 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/7e6b0b81-7dd2-4030-827e-54bbc1a59635-ovn-rundir\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.599638 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/7e6b0b81-7dd2-4030-827e-54bbc1a59635-ovs-rundir\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.600234 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e6b0b81-7dd2-4030-827e-54bbc1a59635-config\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.604281 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e6b0b81-7dd2-4030-827e-54bbc1a59635-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.605209 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e6b0b81-7dd2-4030-827e-54bbc1a59635-combined-ca-bundle\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.628795 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4kgj\" (UniqueName: \"kubernetes.io/projected/7e6b0b81-7dd2-4030-827e-54bbc1a59635-kube-api-access-g4kgj\") pod \"ovn-controller-metrics-nsnv2\" (UID: \"7e6b0b81-7dd2-4030-827e-54bbc1a59635\") " pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.658805 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-4gtkg"] Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.703526 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-config\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.703696 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-dns-svc\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.703863 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.704002 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2zkm\" (UniqueName: \"kubernetes.io/projected/7f9b819c-e185-4a98-8f78-32951d751132-kube-api-access-m2zkm\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.706555 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-dns-svc\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.706703 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.706726 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-config\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.763546 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2zkm\" (UniqueName: \"kubernetes.io/projected/7f9b819c-e185-4a98-8f78-32951d751132-kube-api-access-m2zkm\") pod \"dnsmasq-dns-57d65f699f-4gtkg\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.787614 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-nsnv2" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.789143 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-kmzv6"] Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.802125 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9thnp"] Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.803886 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.806380 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.812800 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9thnp"] Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.905878 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.907528 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-config\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.907623 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgwq4\" (UniqueName: \"kubernetes.io/projected/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-kube-api-access-bgwq4\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.907714 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.907921 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:51 crc kubenswrapper[4744]: I0106 14:56:51.908088 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.009370 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.009533 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.009638 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.009724 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-config\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.009769 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgwq4\" (UniqueName: \"kubernetes.io/projected/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-kube-api-access-bgwq4\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.010626 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.010801 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.010846 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.011073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-config\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.027708 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgwq4\" (UniqueName: \"kubernetes.io/projected/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-kube-api-access-bgwq4\") pod \"dnsmasq-dns-b8fbc5445-9thnp\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.129088 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.187958 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.213823 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tv7b\" (UniqueName: \"kubernetes.io/projected/0d545816-db6b-4953-99a2-716ed95439ac-kube-api-access-4tv7b\") pod \"0d545816-db6b-4953-99a2-716ed95439ac\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.214065 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-config\") pod \"0d545816-db6b-4953-99a2-716ed95439ac\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.214144 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-dns-svc\") pod \"0d545816-db6b-4953-99a2-716ed95439ac\" (UID: \"0d545816-db6b-4953-99a2-716ed95439ac\") " Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.222791 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d545816-db6b-4953-99a2-716ed95439ac-kube-api-access-4tv7b" (OuterVolumeSpecName: "kube-api-access-4tv7b") pod "0d545816-db6b-4953-99a2-716ed95439ac" (UID: "0d545816-db6b-4953-99a2-716ed95439ac"). InnerVolumeSpecName "kube-api-access-4tv7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.262707 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-config" (OuterVolumeSpecName: "config") pod "0d545816-db6b-4953-99a2-716ed95439ac" (UID: "0d545816-db6b-4953-99a2-716ed95439ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.266449 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d545816-db6b-4953-99a2-716ed95439ac" (UID: "0d545816-db6b-4953-99a2-716ed95439ac"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.295701 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-857bd77bfc-rvs27_4b070bab-3e2c-4579-8d78-92ad1f973e62/console/0.log" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.295764 4744 generic.go:334] "Generic (PLEG): container finished" podID="4b070bab-3e2c-4579-8d78-92ad1f973e62" containerID="0f1a7c93518967526038315d0f2e18176895b43b2cfeeb421285a669ffb57ef8" exitCode=2 Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.295842 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-857bd77bfc-rvs27" event={"ID":"4b070bab-3e2c-4579-8d78-92ad1f973e62","Type":"ContainerDied","Data":"0f1a7c93518967526038315d0f2e18176895b43b2cfeeb421285a669ffb57ef8"} Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.301038 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" event={"ID":"0d545816-db6b-4953-99a2-716ed95439ac","Type":"ContainerDied","Data":"d19600dbcdd675a8553b7c7817e8bca4a485c84e3a9d6b450b28326ef2ecc821"} Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.301097 4744 scope.go:117] "RemoveContainer" containerID="2928d12e981531521cd7af0cbf7c185997a971cb2860ae7c53a29e35d7850dd2" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.301118 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-mlnxx" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.316637 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.316663 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d545816-db6b-4953-99a2-716ed95439ac-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.316674 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tv7b\" (UniqueName: \"kubernetes.io/projected/0d545816-db6b-4953-99a2-716ed95439ac-kube-api-access-4tv7b\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:52 crc kubenswrapper[4744]: E0106 14:56:52.321989 4744 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Jan 06 14:56:52 crc kubenswrapper[4744]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/aa886d28-0827-4dae-84bf-a69f196a8933/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Jan 06 14:56:52 crc kubenswrapper[4744]: > podSandboxID="f14cb111de740a1f15eefec48066f285d5995220b1361a81b7119f92a3d25a99" Jan 06 14:56:52 crc kubenswrapper[4744]: E0106 14:56:52.322200 4744 kuberuntime_manager.go:1274] "Unhandled Error" err=< Jan 06 14:56:52 crc kubenswrapper[4744]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4sp5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-cmsvx_openstack(aa886d28-0827-4dae-84bf-a69f196a8933): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/aa886d28-0827-4dae-84bf-a69f196a8933/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Jan 06 14:56:52 crc kubenswrapper[4744]: > logger="UnhandledError" Jan 06 14:56:52 crc kubenswrapper[4744]: E0106 14:56:52.323486 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/aa886d28-0827-4dae-84bf-a69f196a8933/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" podUID="aa886d28-0827-4dae-84bf-a69f196a8933" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.420898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:56:52 crc kubenswrapper[4744]: E0106 14:56:52.421952 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 06 14:56:52 crc kubenswrapper[4744]: E0106 14:56:52.423307 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 06 14:56:52 crc kubenswrapper[4744]: E0106 14:56:52.423359 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift podName:c603cde3-b7b0-4851-99df-2abb0aa6b1c4 nodeName:}" failed. No retries permitted until 2026-01-06 14:57:00.423337617 +0000 UTC m=+1217.050804025 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift") pod "swift-storage-0" (UID: "c603cde3-b7b0-4851-99df-2abb0aa6b1c4") : configmap "swift-ring-files" not found Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.428635 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mlnxx"] Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.449690 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-mlnxx"] Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.772745 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.772821 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.818059 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.826651 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.873040 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.949616 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9thnp"] Jan 06 14:56:52 crc kubenswrapper[4744]: W0106 14:56:52.965766 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd70b136_b28e_4bb2_b85f_e0718fb3ad67.slice/crio-1c45e81f00c94715f0cc1650785e5318c837da0b1bed7b2e87a46f483f6231a0 WatchSource:0}: Error finding container 1c45e81f00c94715f0cc1650785e5318c837da0b1bed7b2e87a46f483f6231a0: Status 404 returned error can't find the container with id 1c45e81f00c94715f0cc1650785e5318c837da0b1bed7b2e87a46f483f6231a0 Jan 06 14:56:52 crc kubenswrapper[4744]: I0106 14:56:52.991776 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-nsnv2"] Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.020280 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-4gtkg"] Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.155146 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-857bd77bfc-rvs27_4b070bab-3e2c-4579-8d78-92ad1f973e62/console/0.log" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.155251 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.198593 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.242848 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-src2w\" (UniqueName: \"kubernetes.io/projected/4b070bab-3e2c-4579-8d78-92ad1f973e62-kube-api-access-src2w\") pod \"4b070bab-3e2c-4579-8d78-92ad1f973e62\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.242924 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-trusted-ca-bundle\") pod \"4b070bab-3e2c-4579-8d78-92ad1f973e62\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.243090 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-oauth-config\") pod \"4b070bab-3e2c-4579-8d78-92ad1f973e62\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.243113 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-oauth-serving-cert\") pod \"4b070bab-3e2c-4579-8d78-92ad1f973e62\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.243150 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-serving-cert\") pod \"4b070bab-3e2c-4579-8d78-92ad1f973e62\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.243231 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-service-ca\") pod \"4b070bab-3e2c-4579-8d78-92ad1f973e62\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.243262 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-config\") pod \"4b070bab-3e2c-4579-8d78-92ad1f973e62\" (UID: \"4b070bab-3e2c-4579-8d78-92ad1f973e62\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.244464 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-config" (OuterVolumeSpecName: "console-config") pod "4b070bab-3e2c-4579-8d78-92ad1f973e62" (UID: "4b070bab-3e2c-4579-8d78-92ad1f973e62"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.244837 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4b070bab-3e2c-4579-8d78-92ad1f973e62" (UID: "4b070bab-3e2c-4579-8d78-92ad1f973e62"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.249659 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4b070bab-3e2c-4579-8d78-92ad1f973e62" (UID: "4b070bab-3e2c-4579-8d78-92ad1f973e62"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.249986 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-service-ca" (OuterVolumeSpecName: "service-ca") pod "4b070bab-3e2c-4579-8d78-92ad1f973e62" (UID: "4b070bab-3e2c-4579-8d78-92ad1f973e62"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.261588 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b070bab-3e2c-4579-8d78-92ad1f973e62-kube-api-access-src2w" (OuterVolumeSpecName: "kube-api-access-src2w") pod "4b070bab-3e2c-4579-8d78-92ad1f973e62" (UID: "4b070bab-3e2c-4579-8d78-92ad1f973e62"). InnerVolumeSpecName "kube-api-access-src2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.269283 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4b070bab-3e2c-4579-8d78-92ad1f973e62" (UID: "4b070bab-3e2c-4579-8d78-92ad1f973e62"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.280420 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4b070bab-3e2c-4579-8d78-92ad1f973e62" (UID: "4b070bab-3e2c-4579-8d78-92ad1f973e62"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.334358 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" event={"ID":"dd70b136-b28e-4bb2-b85f-e0718fb3ad67","Type":"ContainerStarted","Data":"1c45e81f00c94715f0cc1650785e5318c837da0b1bed7b2e87a46f483f6231a0"} Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.345835 4744 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.345874 4744 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.345889 4744 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.345900 4744 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-service-ca\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.345930 4744 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-console-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.345941 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-src2w\" (UniqueName: \"kubernetes.io/projected/4b070bab-3e2c-4579-8d78-92ad1f973e62-kube-api-access-src2w\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.345953 4744 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b070bab-3e2c-4579-8d78-92ad1f973e62-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.351388 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nsnv2" event={"ID":"7e6b0b81-7dd2-4030-827e-54bbc1a59635","Type":"ContainerStarted","Data":"1a28ca3ae41dbc6cdcc2dbaff8ed5a953b17877560a3a31dc4fe78c802d31e48"} Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.364738 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" event={"ID":"61664168-97f1-4e19-920a-d584ea5074b4","Type":"ContainerStarted","Data":"0fec5d709caa635116058224c47efa2a329741d3f1a27fabf15a2a7e26b16c41"} Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.364909 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" podUID="61664168-97f1-4e19-920a-d584ea5074b4" containerName="dnsmasq-dns" containerID="cri-o://0fec5d709caa635116058224c47efa2a329741d3f1a27fabf15a2a7e26b16c41" gracePeriod=10 Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.365007 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.369070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" event={"ID":"7f9b819c-e185-4a98-8f78-32951d751132","Type":"ContainerStarted","Data":"cda5d06447babd92e28e17f18aed2d3b8cb53404c76c1ac07852e79a57282285"} Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.378614 4744 generic.go:334] "Generic (PLEG): container finished" podID="702af028-b407-4750-a0bc-40fc8f144163" containerID="bc5b62258060bf20a1d4f1eb6935143bfd990dbab407532cfdf16ae29ab74621" exitCode=0 Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.378688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerDied","Data":"bc5b62258060bf20a1d4f1eb6935143bfd990dbab407532cfdf16ae29ab74621"} Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.382273 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-857bd77bfc-rvs27_4b070bab-3e2c-4579-8d78-92ad1f973e62/console/0.log" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.382504 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-857bd77bfc-rvs27" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.383226 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-857bd77bfc-rvs27" event={"ID":"4b070bab-3e2c-4579-8d78-92ad1f973e62","Type":"ContainerDied","Data":"74ffafa99f41e0ef114ec6d3df28c1e9201cda09c4332b19f6a5748bd04b96fd"} Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.383280 4744 scope.go:117] "RemoveContainer" containerID="0f1a7c93518967526038315d0f2e18176895b43b2cfeeb421285a669ffb57ef8" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.393975 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.417852 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.459618 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" podStartSLOduration=4.722750661 podStartE2EDuration="10.459528129s" podCreationTimestamp="2026-01-06 14:56:43 +0000 UTC" firstStartedPulling="2026-01-06 14:56:44.371886659 +0000 UTC m=+1200.999352977" lastFinishedPulling="2026-01-06 14:56:50.108664127 +0000 UTC m=+1206.736130445" observedRunningTime="2026-01-06 14:56:53.386646592 +0000 UTC m=+1210.014112910" watchObservedRunningTime="2026-01-06 14:56:53.459528129 +0000 UTC m=+1210.086994447" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.473641 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.494804 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.557279 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-857bd77bfc-rvs27"] Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.570844 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-857bd77bfc-rvs27"] Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.723491 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d545816-db6b-4953-99a2-716ed95439ac" path="/var/lib/kubelet/pods/0d545816-db6b-4953-99a2-716ed95439ac/volumes" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.724317 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b070bab-3e2c-4579-8d78-92ad1f973e62" path="/var/lib/kubelet/pods/4b070bab-3e2c-4579-8d78-92ad1f973e62/volumes" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.862515 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Jan 06 14:56:53 crc kubenswrapper[4744]: E0106 14:56:53.863000 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b070bab-3e2c-4579-8d78-92ad1f973e62" containerName="console" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.863012 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b070bab-3e2c-4579-8d78-92ad1f973e62" containerName="console" Jan 06 14:56:53 crc kubenswrapper[4744]: E0106 14:56:53.863031 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d545816-db6b-4953-99a2-716ed95439ac" containerName="init" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.863038 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d545816-db6b-4953-99a2-716ed95439ac" containerName="init" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.863236 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b070bab-3e2c-4579-8d78-92ad1f973e62" containerName="console" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.863260 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d545816-db6b-4953-99a2-716ed95439ac" containerName="init" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.864292 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.871916 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.872122 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.872263 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.872381 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-24scj" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.884822 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.926961 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.975315 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-config\") pod \"aa886d28-0827-4dae-84bf-a69f196a8933\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.975389 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sp5b\" (UniqueName: \"kubernetes.io/projected/aa886d28-0827-4dae-84bf-a69f196a8933-kube-api-access-4sp5b\") pod \"aa886d28-0827-4dae-84bf-a69f196a8933\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.975432 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-dns-svc\") pod \"aa886d28-0827-4dae-84bf-a69f196a8933\" (UID: \"aa886d28-0827-4dae-84bf-a69f196a8933\") " Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.975955 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73416f0d-1377-4ac4-ab12-c5d54151325b-scripts\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.975988 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/73416f0d-1377-4ac4-ab12-c5d54151325b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.976012 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll7hx\" (UniqueName: \"kubernetes.io/projected/73416f0d-1377-4ac4-ab12-c5d54151325b-kube-api-access-ll7hx\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.976042 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73416f0d-1377-4ac4-ab12-c5d54151325b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.976057 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73416f0d-1377-4ac4-ab12-c5d54151325b-config\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.976138 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/73416f0d-1377-4ac4-ab12-c5d54151325b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.976196 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/73416f0d-1377-4ac4-ab12-c5d54151325b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:53 crc kubenswrapper[4744]: I0106 14:56:53.983218 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa886d28-0827-4dae-84bf-a69f196a8933-kube-api-access-4sp5b" (OuterVolumeSpecName: "kube-api-access-4sp5b") pod "aa886d28-0827-4dae-84bf-a69f196a8933" (UID: "aa886d28-0827-4dae-84bf-a69f196a8933"). InnerVolumeSpecName "kube-api-access-4sp5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.042238 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-config" (OuterVolumeSpecName: "config") pod "aa886d28-0827-4dae-84bf-a69f196a8933" (UID: "aa886d28-0827-4dae-84bf-a69f196a8933"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.046699 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa886d28-0827-4dae-84bf-a69f196a8933" (UID: "aa886d28-0827-4dae-84bf-a69f196a8933"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.078902 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73416f0d-1377-4ac4-ab12-c5d54151325b-scripts\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079070 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/73416f0d-1377-4ac4-ab12-c5d54151325b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079105 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll7hx\" (UniqueName: \"kubernetes.io/projected/73416f0d-1377-4ac4-ab12-c5d54151325b-kube-api-access-ll7hx\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079149 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73416f0d-1377-4ac4-ab12-c5d54151325b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079184 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73416f0d-1377-4ac4-ab12-c5d54151325b-config\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/73416f0d-1377-4ac4-ab12-c5d54151325b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079351 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/73416f0d-1377-4ac4-ab12-c5d54151325b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079451 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079466 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa886d28-0827-4dae-84bf-a69f196a8933-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079477 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sp5b\" (UniqueName: \"kubernetes.io/projected/aa886d28-0827-4dae-84bf-a69f196a8933-kube-api-access-4sp5b\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079873 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/73416f0d-1377-4ac4-ab12-c5d54151325b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.079958 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/73416f0d-1377-4ac4-ab12-c5d54151325b-scripts\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.080757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73416f0d-1377-4ac4-ab12-c5d54151325b-config\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.086717 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/73416f0d-1377-4ac4-ab12-c5d54151325b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.088114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/73416f0d-1377-4ac4-ab12-c5d54151325b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.091305 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73416f0d-1377-4ac4-ab12-c5d54151325b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.096898 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll7hx\" (UniqueName: \"kubernetes.io/projected/73416f0d-1377-4ac4-ab12-c5d54151325b-kube-api-access-ll7hx\") pod \"ovn-northd-0\" (UID: \"73416f0d-1377-4ac4-ab12-c5d54151325b\") " pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.250894 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.424095 4744 generic.go:334] "Generic (PLEG): container finished" podID="7f9b819c-e185-4a98-8f78-32951d751132" containerID="6f242fe9be30286a41017c6e3c8d2750feca44df689bee4dea5898e45b5959cd" exitCode=0 Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.424299 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" event={"ID":"7f9b819c-e185-4a98-8f78-32951d751132","Type":"ContainerDied","Data":"6f242fe9be30286a41017c6e3c8d2750feca44df689bee4dea5898e45b5959cd"} Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.434967 4744 generic.go:334] "Generic (PLEG): container finished" podID="dd70b136-b28e-4bb2-b85f-e0718fb3ad67" containerID="b8e76c4d53d93ae810ab2620fcc4022d458b91aba0cb92e5f7264fbbb909b39b" exitCode=0 Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.435024 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" event={"ID":"dd70b136-b28e-4bb2-b85f-e0718fb3ad67","Type":"ContainerDied","Data":"b8e76c4d53d93ae810ab2620fcc4022d458b91aba0cb92e5f7264fbbb909b39b"} Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.447721 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-nsnv2" event={"ID":"7e6b0b81-7dd2-4030-827e-54bbc1a59635","Type":"ContainerStarted","Data":"0110d10a970ec9718d1a328392ca0d416638004585adc865e4296e874c784e3d"} Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.452056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" event={"ID":"aa886d28-0827-4dae-84bf-a69f196a8933","Type":"ContainerDied","Data":"f14cb111de740a1f15eefec48066f285d5995220b1361a81b7119f92a3d25a99"} Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.452121 4744 scope.go:117] "RemoveContainer" containerID="d7d54a92a83860382f5a94d088a01b8d96650c1308e8c59070bdb14535f45d09" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.452117 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cmsvx" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.454997 4744 generic.go:334] "Generic (PLEG): container finished" podID="61664168-97f1-4e19-920a-d584ea5074b4" containerID="0fec5d709caa635116058224c47efa2a329741d3f1a27fabf15a2a7e26b16c41" exitCode=0 Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.455068 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" event={"ID":"61664168-97f1-4e19-920a-d584ea5074b4","Type":"ContainerDied","Data":"0fec5d709caa635116058224c47efa2a329741d3f1a27fabf15a2a7e26b16c41"} Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.502634 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-nsnv2" podStartSLOduration=3.502613062 podStartE2EDuration="3.502613062s" podCreationTimestamp="2026-01-06 14:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:56:54.476912673 +0000 UTC m=+1211.104378991" watchObservedRunningTime="2026-01-06 14:56:54.502613062 +0000 UTC m=+1211.130079390" Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.549478 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cmsvx"] Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.567822 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cmsvx"] Jan 06 14:56:54 crc kubenswrapper[4744]: I0106 14:56:54.780339 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.741064 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa886d28-0827-4dae-84bf-a69f196a8933" path="/var/lib/kubelet/pods/aa886d28-0827-4dae-84bf-a69f196a8933/volumes" Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.769800 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.834102 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-dns-svc\") pod \"61664168-97f1-4e19-920a-d584ea5074b4\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.834205 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-config\") pod \"61664168-97f1-4e19-920a-d584ea5074b4\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.834779 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ltld\" (UniqueName: \"kubernetes.io/projected/61664168-97f1-4e19-920a-d584ea5074b4-kube-api-access-2ltld\") pod \"61664168-97f1-4e19-920a-d584ea5074b4\" (UID: \"61664168-97f1-4e19-920a-d584ea5074b4\") " Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.838687 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61664168-97f1-4e19-920a-d584ea5074b4-kube-api-access-2ltld" (OuterVolumeSpecName: "kube-api-access-2ltld") pod "61664168-97f1-4e19-920a-d584ea5074b4" (UID: "61664168-97f1-4e19-920a-d584ea5074b4"). InnerVolumeSpecName "kube-api-access-2ltld". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.897870 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-config" (OuterVolumeSpecName: "config") pod "61664168-97f1-4e19-920a-d584ea5074b4" (UID: "61664168-97f1-4e19-920a-d584ea5074b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.925985 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "61664168-97f1-4e19-920a-d584ea5074b4" (UID: "61664168-97f1-4e19-920a-d584ea5074b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.938374 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.938412 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61664168-97f1-4e19-920a-d584ea5074b4-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:55 crc kubenswrapper[4744]: I0106 14:56:55.938429 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ltld\" (UniqueName: \"kubernetes.io/projected/61664168-97f1-4e19-920a-d584ea5074b4-kube-api-access-2ltld\") on node \"crc\" DevicePath \"\"" Jan 06 14:56:56 crc kubenswrapper[4744]: I0106 14:56:56.481039 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" event={"ID":"61664168-97f1-4e19-920a-d584ea5074b4","Type":"ContainerDied","Data":"d809eaf79b810e50bfb6383d5e296353530653c9a0150b286a6aaadb2b415ee2"} Jan 06 14:56:56 crc kubenswrapper[4744]: I0106 14:56:56.481476 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-kmzv6" Jan 06 14:56:56 crc kubenswrapper[4744]: I0106 14:56:56.489508 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"73416f0d-1377-4ac4-ab12-c5d54151325b","Type":"ContainerStarted","Data":"47329f8e34858c4fecae94ee72aa2facd6d0fee7f11e611f8c6ac06c80f26d53"} Jan 06 14:56:56 crc kubenswrapper[4744]: I0106 14:56:56.530011 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-kmzv6"] Jan 06 14:56:56 crc kubenswrapper[4744]: I0106 14:56:56.540311 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-kmzv6"] Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.427068 4744 scope.go:117] "RemoveContainer" containerID="0fec5d709caa635116058224c47efa2a329741d3f1a27fabf15a2a7e26b16c41" Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.511465 4744 generic.go:334] "Generic (PLEG): container finished" podID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerID="23858f13f4cc753d4995c5de460b682039086270f2029befd5f5afb9b8976ea7" exitCode=0 Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.511557 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4","Type":"ContainerDied","Data":"23858f13f4cc753d4995c5de460b682039086270f2029befd5f5afb9b8976ea7"} Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.516363 4744 generic.go:334] "Generic (PLEG): container finished" podID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerID="7bdc826178224ec8f48dc20281097da9db5ad81ce52b8030850b40e45948d25b" exitCode=0 Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.516435 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"d030be7f-cfb2-4335-a2f8-7da8f7430b17","Type":"ContainerDied","Data":"7bdc826178224ec8f48dc20281097da9db5ad81ce52b8030850b40e45948d25b"} Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.523736 4744 generic.go:334] "Generic (PLEG): container finished" podID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerID="b568e5b0f4450110f1563bd5f836d3c37170139a7c49b77607d38119a536ea25" exitCode=0 Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.523876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff","Type":"ContainerDied","Data":"b568e5b0f4450110f1563bd5f836d3c37170139a7c49b77607d38119a536ea25"} Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.537738 4744 generic.go:334] "Generic (PLEG): container finished" podID="bef00f62-0142-4a72-9849-12246c13d004" containerID="738a503b4e10e99e3abe518b1919d9ce5bac03ecfcbea228a9694a955f0cf80d" exitCode=0 Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.537808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"bef00f62-0142-4a72-9849-12246c13d004","Type":"ContainerDied","Data":"738a503b4e10e99e3abe518b1919d9ce5bac03ecfcbea228a9694a955f0cf80d"} Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.733435 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61664168-97f1-4e19-920a-d584ea5074b4" path="/var/lib/kubelet/pods/61664168-97f1-4e19-920a-d584ea5074b4/volumes" Jan 06 14:56:57 crc kubenswrapper[4744]: I0106 14:56:57.839341 4744 scope.go:117] "RemoveContainer" containerID="814b24f3776dc8599ba6be2125e8fc7c87a8bafce5ef853eab7d926a7d0fa25f" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.534475 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-xf2lt"] Jan 06 14:56:58 crc kubenswrapper[4744]: E0106 14:56:58.534901 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa886d28-0827-4dae-84bf-a69f196a8933" containerName="init" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.534913 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa886d28-0827-4dae-84bf-a69f196a8933" containerName="init" Jan 06 14:56:58 crc kubenswrapper[4744]: E0106 14:56:58.534927 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61664168-97f1-4e19-920a-d584ea5074b4" containerName="dnsmasq-dns" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.534932 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="61664168-97f1-4e19-920a-d584ea5074b4" containerName="dnsmasq-dns" Jan 06 14:56:58 crc kubenswrapper[4744]: E0106 14:56:58.534945 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61664168-97f1-4e19-920a-d584ea5074b4" containerName="init" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.534952 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="61664168-97f1-4e19-920a-d584ea5074b4" containerName="init" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.535143 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa886d28-0827-4dae-84bf-a69f196a8933" containerName="init" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.535171 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="61664168-97f1-4e19-920a-d584ea5074b4" containerName="dnsmasq-dns" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.535960 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xf2lt" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.541555 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.553471 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fq864" event={"ID":"82b9e690-d1d7-473d-99d1-20be1130dfd3","Type":"ContainerStarted","Data":"1798a88497d35559965f70f8539e90b4e5019f317139ea4a0a68b98930885e8f"} Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.557626 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4","Type":"ContainerStarted","Data":"e8e69d68f403efbf1a5c144e3e5408e06f7ea892efbde73e923d23e93353ab57"} Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.558034 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.572402 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" event={"ID":"7f9b819c-e185-4a98-8f78-32951d751132","Type":"ContainerStarted","Data":"5fc6fe38c52a96699333bf19a79116c1d0bd45a68663e3959bf32831e6ac240f"} Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.573374 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-xf2lt"] Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.573596 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.585666 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"d030be7f-cfb2-4335-a2f8-7da8f7430b17","Type":"ContainerStarted","Data":"595b0c188fa47a16c899d11b4a1a07707b58c92ed185b449bf699f6384597f4e"} Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.587250 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.592283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"bef00f62-0142-4a72-9849-12246c13d004","Type":"ContainerStarted","Data":"98ff97c26acef954ef996e4aacd4e7dec375b11833016c6e200cd86dd7f206d4"} Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.592598 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.596875 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" event={"ID":"dd70b136-b28e-4bb2-b85f-e0718fb3ad67","Type":"ContainerStarted","Data":"2da3e48201e6ebeb7f81d43329815c196b5f93a2c494e874c63ee3331711065d"} Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.597990 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.615042 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.291226556 podStartE2EDuration="1m1.614937411s" podCreationTimestamp="2026-01-06 14:55:57 +0000 UTC" firstStartedPulling="2026-01-06 14:56:00.254802192 +0000 UTC m=+1156.882268500" lastFinishedPulling="2026-01-06 14:56:23.578513037 +0000 UTC m=+1180.205979355" observedRunningTime="2026-01-06 14:56:58.607839553 +0000 UTC m=+1215.235305881" watchObservedRunningTime="2026-01-06 14:56:58.614937411 +0000 UTC m=+1215.242403739" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.637978 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-fq864" podStartSLOduration=6.36314693 podStartE2EDuration="13.637871637s" podCreationTimestamp="2026-01-06 14:56:45 +0000 UTC" firstStartedPulling="2026-01-06 14:56:50.266355497 +0000 UTC m=+1206.893821815" lastFinishedPulling="2026-01-06 14:56:57.541080204 +0000 UTC m=+1214.168546522" observedRunningTime="2026-01-06 14:56:58.634101338 +0000 UTC m=+1215.261567656" watchObservedRunningTime="2026-01-06 14:56:58.637871637 +0000 UTC m=+1215.265337965" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.676194 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" podStartSLOduration=7.676109109 podStartE2EDuration="7.676109109s" podCreationTimestamp="2026-01-06 14:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:56:58.664154523 +0000 UTC m=+1215.291620851" watchObservedRunningTime="2026-01-06 14:56:58.676109109 +0000 UTC m=+1215.303575427" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.715066 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=38.894689519 podStartE2EDuration="1m2.715035948s" podCreationTimestamp="2026-01-06 14:55:56 +0000 UTC" firstStartedPulling="2026-01-06 14:55:59.669205596 +0000 UTC m=+1156.296671914" lastFinishedPulling="2026-01-06 14:56:23.489552025 +0000 UTC m=+1180.117018343" observedRunningTime="2026-01-06 14:56:58.694757622 +0000 UTC m=+1215.322223950" watchObservedRunningTime="2026-01-06 14:56:58.715035948 +0000 UTC m=+1215.342502286" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.715652 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khf8h\" (UniqueName: \"kubernetes.io/projected/ac516648-534b-4d75-ae6e-d646c0d268e8-kube-api-access-khf8h\") pod \"root-account-create-update-xf2lt\" (UID: \"ac516648-534b-4d75-ae6e-d646c0d268e8\") " pod="openstack/root-account-create-update-xf2lt" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.716087 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac516648-534b-4d75-ae6e-d646c0d268e8-operator-scripts\") pod \"root-account-create-update-xf2lt\" (UID: \"ac516648-534b-4d75-ae6e-d646c0d268e8\") " pod="openstack/root-account-create-update-xf2lt" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.753732 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" podStartSLOduration=7.75366185 podStartE2EDuration="7.75366185s" podCreationTimestamp="2026-01-06 14:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:56:58.730520868 +0000 UTC m=+1215.357987186" watchObservedRunningTime="2026-01-06 14:56:58.75366185 +0000 UTC m=+1215.381128178" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.774859 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=38.861721307 podStartE2EDuration="1m2.774837679s" podCreationTimestamp="2026-01-06 14:55:56 +0000 UTC" firstStartedPulling="2026-01-06 14:55:59.69998761 +0000 UTC m=+1156.327453928" lastFinishedPulling="2026-01-06 14:56:23.613103982 +0000 UTC m=+1180.240570300" observedRunningTime="2026-01-06 14:56:58.759115554 +0000 UTC m=+1215.386581892" watchObservedRunningTime="2026-01-06 14:56:58.774837679 +0000 UTC m=+1215.402303997" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.818344 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac516648-534b-4d75-ae6e-d646c0d268e8-operator-scripts\") pod \"root-account-create-update-xf2lt\" (UID: \"ac516648-534b-4d75-ae6e-d646c0d268e8\") " pod="openstack/root-account-create-update-xf2lt" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.818724 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khf8h\" (UniqueName: \"kubernetes.io/projected/ac516648-534b-4d75-ae6e-d646c0d268e8-kube-api-access-khf8h\") pod \"root-account-create-update-xf2lt\" (UID: \"ac516648-534b-4d75-ae6e-d646c0d268e8\") " pod="openstack/root-account-create-update-xf2lt" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.820993 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac516648-534b-4d75-ae6e-d646c0d268e8-operator-scripts\") pod \"root-account-create-update-xf2lt\" (UID: \"ac516648-534b-4d75-ae6e-d646c0d268e8\") " pod="openstack/root-account-create-update-xf2lt" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.838957 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khf8h\" (UniqueName: \"kubernetes.io/projected/ac516648-534b-4d75-ae6e-d646c0d268e8-kube-api-access-khf8h\") pod \"root-account-create-update-xf2lt\" (UID: \"ac516648-534b-4d75-ae6e-d646c0d268e8\") " pod="openstack/root-account-create-update-xf2lt" Jan 06 14:56:58 crc kubenswrapper[4744]: I0106 14:56:58.893572 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xf2lt" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:56:59.657455 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"73416f0d-1377-4ac4-ab12-c5d54151325b","Type":"ContainerStarted","Data":"e61a33f99b8ecd1cb77ae6f28c1efc1b718c9261e1a0d078258c51aedd5e0fde"} Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:56:59.665891 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff","Type":"ContainerStarted","Data":"c6f4211a18370302a96e5b9028a1ccfd0b2c832ac6b5ed9dbcf6df60c491d43f"} Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:56:59.693772 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.781224944 podStartE2EDuration="1m3.69375585s" podCreationTimestamp="2026-01-06 14:55:56 +0000 UTC" firstStartedPulling="2026-01-06 14:55:59.699439226 +0000 UTC m=+1156.326905544" lastFinishedPulling="2026-01-06 14:56:23.611970132 +0000 UTC m=+1180.239436450" observedRunningTime="2026-01-06 14:56:59.690496154 +0000 UTC m=+1216.317962482" watchObservedRunningTime="2026-01-06 14:56:59.69375585 +0000 UTC m=+1216.321222168" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:00.464014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:57:01 crc kubenswrapper[4744]: E0106 14:57:00.464247 4744 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 06 14:57:01 crc kubenswrapper[4744]: E0106 14:57:00.464279 4744 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 06 14:57:01 crc kubenswrapper[4744]: E0106 14:57:00.464341 4744 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift podName:c603cde3-b7b0-4851-99df-2abb0aa6b1c4 nodeName:}" failed. No retries permitted until 2026-01-06 14:57:16.464319378 +0000 UTC m=+1233.091785696 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift") pod "swift-storage-0" (UID: "c603cde3-b7b0-4851-99df-2abb0aa6b1c4") : configmap "swift-ring-files" not found Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:00.948010 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8d3a-account-create-update-fhjdm"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:00.950744 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:00.953093 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:00.960132 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8d3a-account-create-update-fhjdm"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:00.994444 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-2rcd6"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:00.998509 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.001699 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2rcd6"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.077880 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzrf9\" (UniqueName: \"kubernetes.io/projected/d7e8cff6-96c6-4fee-99ca-1f77403080b4-kube-api-access-kzrf9\") pod \"keystone-8d3a-account-create-update-fhjdm\" (UID: \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\") " pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.078378 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8cff6-96c6-4fee-99ca-1f77403080b4-operator-scripts\") pod \"keystone-8d3a-account-create-update-fhjdm\" (UID: \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\") " pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.078523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl2z9\" (UniqueName: \"kubernetes.io/projected/bf899aa6-5d99-482f-b493-56d7de87e44c-kube-api-access-rl2z9\") pod \"keystone-db-create-2rcd6\" (UID: \"bf899aa6-5d99-482f-b493-56d7de87e44c\") " pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.078854 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf899aa6-5d99-482f-b493-56d7de87e44c-operator-scripts\") pod \"keystone-db-create-2rcd6\" (UID: \"bf899aa6-5d99-482f-b493-56d7de87e44c\") " pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.117517 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-thfs7"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.119257 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-thfs7" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.127504 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-thfs7"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.184053 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzrf9\" (UniqueName: \"kubernetes.io/projected/d7e8cff6-96c6-4fee-99ca-1f77403080b4-kube-api-access-kzrf9\") pod \"keystone-8d3a-account-create-update-fhjdm\" (UID: \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\") " pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.184490 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8cff6-96c6-4fee-99ca-1f77403080b4-operator-scripts\") pod \"keystone-8d3a-account-create-update-fhjdm\" (UID: \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\") " pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.184549 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl2z9\" (UniqueName: \"kubernetes.io/projected/bf899aa6-5d99-482f-b493-56d7de87e44c-kube-api-access-rl2z9\") pod \"keystone-db-create-2rcd6\" (UID: \"bf899aa6-5d99-482f-b493-56d7de87e44c\") " pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.184600 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d755b346-e89a-4ae8-a162-4fc6f8c28797-operator-scripts\") pod \"placement-db-create-thfs7\" (UID: \"d755b346-e89a-4ae8-a162-4fc6f8c28797\") " pod="openstack/placement-db-create-thfs7" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.184623 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdrpq\" (UniqueName: \"kubernetes.io/projected/d755b346-e89a-4ae8-a162-4fc6f8c28797-kube-api-access-sdrpq\") pod \"placement-db-create-thfs7\" (UID: \"d755b346-e89a-4ae8-a162-4fc6f8c28797\") " pod="openstack/placement-db-create-thfs7" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.184718 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf899aa6-5d99-482f-b493-56d7de87e44c-operator-scripts\") pod \"keystone-db-create-2rcd6\" (UID: \"bf899aa6-5d99-482f-b493-56d7de87e44c\") " pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.185646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf899aa6-5d99-482f-b493-56d7de87e44c-operator-scripts\") pod \"keystone-db-create-2rcd6\" (UID: \"bf899aa6-5d99-482f-b493-56d7de87e44c\") " pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.187265 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8cff6-96c6-4fee-99ca-1f77403080b4-operator-scripts\") pod \"keystone-8d3a-account-create-update-fhjdm\" (UID: \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\") " pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.209632 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzrf9\" (UniqueName: \"kubernetes.io/projected/d7e8cff6-96c6-4fee-99ca-1f77403080b4-kube-api-access-kzrf9\") pod \"keystone-8d3a-account-create-update-fhjdm\" (UID: \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\") " pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.238837 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl2z9\" (UniqueName: \"kubernetes.io/projected/bf899aa6-5d99-482f-b493-56d7de87e44c-kube-api-access-rl2z9\") pod \"keystone-db-create-2rcd6\" (UID: \"bf899aa6-5d99-482f-b493-56d7de87e44c\") " pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.279697 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.286761 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d755b346-e89a-4ae8-a162-4fc6f8c28797-operator-scripts\") pod \"placement-db-create-thfs7\" (UID: \"d755b346-e89a-4ae8-a162-4fc6f8c28797\") " pod="openstack/placement-db-create-thfs7" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.286795 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdrpq\" (UniqueName: \"kubernetes.io/projected/d755b346-e89a-4ae8-a162-4fc6f8c28797-kube-api-access-sdrpq\") pod \"placement-db-create-thfs7\" (UID: \"d755b346-e89a-4ae8-a162-4fc6f8c28797\") " pod="openstack/placement-db-create-thfs7" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.287610 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d755b346-e89a-4ae8-a162-4fc6f8c28797-operator-scripts\") pod \"placement-db-create-thfs7\" (UID: \"d755b346-e89a-4ae8-a162-4fc6f8c28797\") " pod="openstack/placement-db-create-thfs7" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.325986 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdrpq\" (UniqueName: \"kubernetes.io/projected/d755b346-e89a-4ae8-a162-4fc6f8c28797-kube-api-access-sdrpq\") pod \"placement-db-create-thfs7\" (UID: \"d755b346-e89a-4ae8-a162-4fc6f8c28797\") " pod="openstack/placement-db-create-thfs7" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.327733 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.387909 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-79d0-account-create-update-w29fr"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.401266 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.408566 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.437043 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-79d0-account-create-update-w29fr"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.451558 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-thfs7" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.504729 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4136bc47-2e87-4af7-acf9-23a36c37269d-operator-scripts\") pod \"placement-79d0-account-create-update-w29fr\" (UID: \"4136bc47-2e87-4af7-acf9-23a36c37269d\") " pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.505469 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6cwd\" (UniqueName: \"kubernetes.io/projected/4136bc47-2e87-4af7-acf9-23a36c37269d-kube-api-access-w6cwd\") pod \"placement-79d0-account-create-update-w29fr\" (UID: \"4136bc47-2e87-4af7-acf9-23a36c37269d\") " pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.541781 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-bv5rj"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.543044 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.578944 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bv5rj"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.608372 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4136bc47-2e87-4af7-acf9-23a36c37269d-operator-scripts\") pod \"placement-79d0-account-create-update-w29fr\" (UID: \"4136bc47-2e87-4af7-acf9-23a36c37269d\") " pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.608449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f410252c-e5a3-4d42-ba06-febebef5a30a-operator-scripts\") pod \"glance-db-create-bv5rj\" (UID: \"f410252c-e5a3-4d42-ba06-febebef5a30a\") " pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.608571 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6cwd\" (UniqueName: \"kubernetes.io/projected/4136bc47-2e87-4af7-acf9-23a36c37269d-kube-api-access-w6cwd\") pod \"placement-79d0-account-create-update-w29fr\" (UID: \"4136bc47-2e87-4af7-acf9-23a36c37269d\") " pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.608621 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq8m4\" (UniqueName: \"kubernetes.io/projected/f410252c-e5a3-4d42-ba06-febebef5a30a-kube-api-access-sq8m4\") pod \"glance-db-create-bv5rj\" (UID: \"f410252c-e5a3-4d42-ba06-febebef5a30a\") " pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.609399 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4136bc47-2e87-4af7-acf9-23a36c37269d-operator-scripts\") pod \"placement-79d0-account-create-update-w29fr\" (UID: \"4136bc47-2e87-4af7-acf9-23a36c37269d\") " pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.649769 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6cwd\" (UniqueName: \"kubernetes.io/projected/4136bc47-2e87-4af7-acf9-23a36c37269d-kube-api-access-w6cwd\") pod \"placement-79d0-account-create-update-w29fr\" (UID: \"4136bc47-2e87-4af7-acf9-23a36c37269d\") " pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.656703 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-f0f3-account-create-update-h8jq2"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.657964 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.660415 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.733494 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f410252c-e5a3-4d42-ba06-febebef5a30a-operator-scripts\") pod \"glance-db-create-bv5rj\" (UID: \"f410252c-e5a3-4d42-ba06-febebef5a30a\") " pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.733569 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nczl5\" (UniqueName: \"kubernetes.io/projected/de589768-1fff-4511-a85e-cbb4e8cb30ed-kube-api-access-nczl5\") pod \"glance-f0f3-account-create-update-h8jq2\" (UID: \"de589768-1fff-4511-a85e-cbb4e8cb30ed\") " pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.733681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq8m4\" (UniqueName: \"kubernetes.io/projected/f410252c-e5a3-4d42-ba06-febebef5a30a-kube-api-access-sq8m4\") pod \"glance-db-create-bv5rj\" (UID: \"f410252c-e5a3-4d42-ba06-febebef5a30a\") " pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.733702 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de589768-1fff-4511-a85e-cbb4e8cb30ed-operator-scripts\") pod \"glance-f0f3-account-create-update-h8jq2\" (UID: \"de589768-1fff-4511-a85e-cbb4e8cb30ed\") " pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.734384 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f410252c-e5a3-4d42-ba06-febebef5a30a-operator-scripts\") pod \"glance-db-create-bv5rj\" (UID: \"f410252c-e5a3-4d42-ba06-febebef5a30a\") " pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.759484 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.767527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq8m4\" (UniqueName: \"kubernetes.io/projected/f410252c-e5a3-4d42-ba06-febebef5a30a-kube-api-access-sq8m4\") pod \"glance-db-create-bv5rj\" (UID: \"f410252c-e5a3-4d42-ba06-febebef5a30a\") " pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.837311 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de589768-1fff-4511-a85e-cbb4e8cb30ed-operator-scripts\") pod \"glance-f0f3-account-create-update-h8jq2\" (UID: \"de589768-1fff-4511-a85e-cbb4e8cb30ed\") " pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.837478 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nczl5\" (UniqueName: \"kubernetes.io/projected/de589768-1fff-4511-a85e-cbb4e8cb30ed-kube-api-access-nczl5\") pod \"glance-f0f3-account-create-update-h8jq2\" (UID: \"de589768-1fff-4511-a85e-cbb4e8cb30ed\") " pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.838638 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de589768-1fff-4511-a85e-cbb4e8cb30ed-operator-scripts\") pod \"glance-f0f3-account-create-update-h8jq2\" (UID: \"de589768-1fff-4511-a85e-cbb4e8cb30ed\") " pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.859487 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nczl5\" (UniqueName: \"kubernetes.io/projected/de589768-1fff-4511-a85e-cbb4e8cb30ed-kube-api-access-nczl5\") pod \"glance-f0f3-account-create-update-h8jq2\" (UID: \"de589768-1fff-4511-a85e-cbb4e8cb30ed\") " pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.876469 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f0f3-account-create-update-h8jq2"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.876525 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-xf2lt"] Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.882995 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:01 crc kubenswrapper[4744]: I0106 14:57:01.886347 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.276293 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-x5r7t"] Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.278705 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.294793 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-x5r7t"] Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.389919 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqbfs\" (UniqueName: \"kubernetes.io/projected/c10fd027-3c56-4a1d-a16f-de6b641a82f4-kube-api-access-vqbfs\") pod \"mysqld-exporter-openstack-db-create-x5r7t\" (UID: \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\") " pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.390085 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c10fd027-3c56-4a1d-a16f-de6b641a82f4-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-x5r7t\" (UID: \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\") " pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.428517 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.438943 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-6c8a-account-create-update-7b4pf"] Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.440463 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.443830 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.468715 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-6c8a-account-create-update-7b4pf"] Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.491718 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c10fd027-3c56-4a1d-a16f-de6b641a82f4-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-x5r7t\" (UID: \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\") " pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.491835 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99zhz\" (UniqueName: \"kubernetes.io/projected/218cb457-4148-4612-bbf6-f8240c136d3b-kube-api-access-99zhz\") pod \"mysqld-exporter-6c8a-account-create-update-7b4pf\" (UID: \"218cb457-4148-4612-bbf6-f8240c136d3b\") " pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.491868 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqbfs\" (UniqueName: \"kubernetes.io/projected/c10fd027-3c56-4a1d-a16f-de6b641a82f4-kube-api-access-vqbfs\") pod \"mysqld-exporter-openstack-db-create-x5r7t\" (UID: \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\") " pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.491922 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218cb457-4148-4612-bbf6-f8240c136d3b-operator-scripts\") pod \"mysqld-exporter-6c8a-account-create-update-7b4pf\" (UID: \"218cb457-4148-4612-bbf6-f8240c136d3b\") " pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.493382 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c10fd027-3c56-4a1d-a16f-de6b641a82f4-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-x5r7t\" (UID: \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\") " pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.520763 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqbfs\" (UniqueName: \"kubernetes.io/projected/c10fd027-3c56-4a1d-a16f-de6b641a82f4-kube-api-access-vqbfs\") pod \"mysqld-exporter-openstack-db-create-x5r7t\" (UID: \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\") " pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.593946 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99zhz\" (UniqueName: \"kubernetes.io/projected/218cb457-4148-4612-bbf6-f8240c136d3b-kube-api-access-99zhz\") pod \"mysqld-exporter-6c8a-account-create-update-7b4pf\" (UID: \"218cb457-4148-4612-bbf6-f8240c136d3b\") " pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.594235 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218cb457-4148-4612-bbf6-f8240c136d3b-operator-scripts\") pod \"mysqld-exporter-6c8a-account-create-update-7b4pf\" (UID: \"218cb457-4148-4612-bbf6-f8240c136d3b\") " pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.594854 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218cb457-4148-4612-bbf6-f8240c136d3b-operator-scripts\") pod \"mysqld-exporter-6c8a-account-create-update-7b4pf\" (UID: \"218cb457-4148-4612-bbf6-f8240c136d3b\") " pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.603419 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.612027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99zhz\" (UniqueName: \"kubernetes.io/projected/218cb457-4148-4612-bbf6-f8240c136d3b-kube-api-access-99zhz\") pod \"mysqld-exporter-6c8a-account-create-update-7b4pf\" (UID: \"218cb457-4148-4612-bbf6-f8240c136d3b\") " pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:03 crc kubenswrapper[4744]: I0106 14:57:03.761903 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:06 crc kubenswrapper[4744]: I0106 14:57:06.908369 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:57:07 crc kubenswrapper[4744]: I0106 14:57:07.131350 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:57:07 crc kubenswrapper[4744]: I0106 14:57:07.190438 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-4gtkg"] Jan 06 14:57:07 crc kubenswrapper[4744]: I0106 14:57:07.797127 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" podUID="7f9b819c-e185-4a98-8f78-32951d751132" containerName="dnsmasq-dns" containerID="cri-o://5fc6fe38c52a96699333bf19a79116c1d0bd45a68663e3959bf32831e6ac240f" gracePeriod=10 Jan 06 14:57:07 crc kubenswrapper[4744]: I0106 14:57:07.857612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xf2lt" event={"ID":"ac516648-534b-4d75-ae6e-d646c0d268e8","Type":"ContainerStarted","Data":"ca190b206102e15d2d99e1edf1d9d8c489c28099aec63af944b3c2030f6a806e"} Jan 06 14:57:07 crc kubenswrapper[4744]: I0106 14:57:07.918619 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-r4r76" podUID="7f86f187-37db-488e-b6f6-22a5becacecd" containerName="ovn-controller" probeResult="failure" output=< Jan 06 14:57:07 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 06 14:57:07 crc kubenswrapper[4744]: > Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.394313 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.407831 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.128:5671: connect: connection refused" Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.408172 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.128:5671: connect: connection refused" Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.471525 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-2rcd6"] Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.725527 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.757557 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.819055 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xf2lt" event={"ID":"ac516648-534b-4d75-ae6e-d646c0d268e8","Type":"ContainerStarted","Data":"c7766c02ae0bac5d27289481b8d90064c15e4b93c06f99ef0a12a9699f8225a2"} Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.835118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"73416f0d-1377-4ac4-ab12-c5d54151325b","Type":"ContainerStarted","Data":"81629b4d43a3c3ad243502d1b947d1ef35456dd76fba16492c62463181ad5406"} Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.835272 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.839604 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2rcd6" event={"ID":"bf899aa6-5d99-482f-b493-56d7de87e44c","Type":"ContainerStarted","Data":"367e85e3fef5ea359c3a9b5ae21d2a29735a77abfcfecce5f5849eb7d383083a"} Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.844643 4744 generic.go:334] "Generic (PLEG): container finished" podID="7f9b819c-e185-4a98-8f78-32951d751132" containerID="5fc6fe38c52a96699333bf19a79116c1d0bd45a68663e3959bf32831e6ac240f" exitCode=0 Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.844704 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" event={"ID":"7f9b819c-e185-4a98-8f78-32951d751132","Type":"ContainerDied","Data":"5fc6fe38c52a96699333bf19a79116c1d0bd45a68663e3959bf32831e6ac240f"} Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.851921 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-xf2lt" podStartSLOduration=10.851897583 podStartE2EDuration="10.851897583s" podCreationTimestamp="2026-01-06 14:56:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:08.848477482 +0000 UTC m=+1225.475943810" watchObservedRunningTime="2026-01-06 14:57:08.851897583 +0000 UTC m=+1225.479363901" Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.859135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerStarted","Data":"e8690c660760c008bf10dfaba0798d89793d116fd6a52d58128117656d2a976b"} Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.883183 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=12.730064627 podStartE2EDuration="15.883130329s" podCreationTimestamp="2026-01-06 14:56:53 +0000 UTC" firstStartedPulling="2026-01-06 14:56:55.627636443 +0000 UTC m=+1212.255102761" lastFinishedPulling="2026-01-06 14:56:58.780702145 +0000 UTC m=+1215.408168463" observedRunningTime="2026-01-06 14:57:08.874682295 +0000 UTC m=+1225.502148623" watchObservedRunningTime="2026-01-06 14:57:08.883130329 +0000 UTC m=+1225.510596647" Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.907578 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-2rcd6" podStartSLOduration=8.907556415 podStartE2EDuration="8.907556415s" podCreationTimestamp="2026-01-06 14:57:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:08.904298019 +0000 UTC m=+1225.531764347" watchObservedRunningTime="2026-01-06 14:57:08.907556415 +0000 UTC m=+1225.535022733" Jan 06 14:57:08 crc kubenswrapper[4744]: I0106 14:57:08.950981 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-x5r7t"] Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.065570 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.117107 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-thfs7"] Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.154491 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-6c8a-account-create-update-7b4pf"] Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.185308 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bv5rj"] Jan 06 14:57:09 crc kubenswrapper[4744]: W0106 14:57:09.193011 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd755b346_e89a_4ae8_a162_4fc6f8c28797.slice/crio-8c500d01712b6c097d0eab8a9c6f3e261342b7437faf36f2da421338175701a8 WatchSource:0}: Error finding container 8c500d01712b6c097d0eab8a9c6f3e261342b7437faf36f2da421338175701a8: Status 404 returned error can't find the container with id 8c500d01712b6c097d0eab8a9c6f3e261342b7437faf36f2da421338175701a8 Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.326418 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-79d0-account-create-update-w29fr"] Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.344664 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8d3a-account-create-update-fhjdm"] Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.397559 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.398355 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f0f3-account-create-update-h8jq2"] Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.461599 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.639659 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2zkm\" (UniqueName: \"kubernetes.io/projected/7f9b819c-e185-4a98-8f78-32951d751132-kube-api-access-m2zkm\") pod \"7f9b819c-e185-4a98-8f78-32951d751132\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.639800 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-ovsdbserver-nb\") pod \"7f9b819c-e185-4a98-8f78-32951d751132\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.639963 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-dns-svc\") pod \"7f9b819c-e185-4a98-8f78-32951d751132\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.640055 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-config\") pod \"7f9b819c-e185-4a98-8f78-32951d751132\" (UID: \"7f9b819c-e185-4a98-8f78-32951d751132\") " Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.675915 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f9b819c-e185-4a98-8f78-32951d751132-kube-api-access-m2zkm" (OuterVolumeSpecName: "kube-api-access-m2zkm") pod "7f9b819c-e185-4a98-8f78-32951d751132" (UID: "7f9b819c-e185-4a98-8f78-32951d751132"). InnerVolumeSpecName "kube-api-access-m2zkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.746205 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2zkm\" (UniqueName: \"kubernetes.io/projected/7f9b819c-e185-4a98-8f78-32951d751132-kube-api-access-m2zkm\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.873510 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" event={"ID":"c10fd027-3c56-4a1d-a16f-de6b641a82f4","Type":"ContainerStarted","Data":"17137168f6ccfc7e674b4bccba59ed5e27bb7941d029088f9ceaa674b5014cdd"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.876861 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" event={"ID":"7f9b819c-e185-4a98-8f78-32951d751132","Type":"ContainerDied","Data":"cda5d06447babd92e28e17f18aed2d3b8cb53404c76c1ac07852e79a57282285"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.877320 4744 scope.go:117] "RemoveContainer" containerID="5fc6fe38c52a96699333bf19a79116c1d0bd45a68663e3959bf32831e6ac240f" Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.876935 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-4gtkg" Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.878495 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8d3a-account-create-update-fhjdm" event={"ID":"d7e8cff6-96c6-4fee-99ca-1f77403080b4","Type":"ContainerStarted","Data":"096cdbf6781798dc9ba817cf19711b776802cde071ad51c5c9bce9e21237ca71"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.880030 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f0f3-account-create-update-h8jq2" event={"ID":"de589768-1fff-4511-a85e-cbb4e8cb30ed","Type":"ContainerStarted","Data":"3b38be9543a2e95b2af32a5b267ab930ed4bab03725e3a733fae1a58beb14d17"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.881035 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bv5rj" event={"ID":"f410252c-e5a3-4d42-ba06-febebef5a30a","Type":"ContainerStarted","Data":"e1f1bf5fdd3ae4f8f066081bc6550d7988bf58f23f9a0163010cd549c7b70cdd"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.882351 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-thfs7" event={"ID":"d755b346-e89a-4ae8-a162-4fc6f8c28797","Type":"ContainerStarted","Data":"8c500d01712b6c097d0eab8a9c6f3e261342b7437faf36f2da421338175701a8"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.884426 4744 generic.go:334] "Generic (PLEG): container finished" podID="ac516648-534b-4d75-ae6e-d646c0d268e8" containerID="c7766c02ae0bac5d27289481b8d90064c15e4b93c06f99ef0a12a9699f8225a2" exitCode=0 Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.884467 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xf2lt" event={"ID":"ac516648-534b-4d75-ae6e-d646c0d268e8","Type":"ContainerDied","Data":"c7766c02ae0bac5d27289481b8d90064c15e4b93c06f99ef0a12a9699f8225a2"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.886528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" event={"ID":"218cb457-4148-4612-bbf6-f8240c136d3b","Type":"ContainerStarted","Data":"dfc823a84423cdf91639caed00468369ea80f1a2deb68af5b189b685d2ad57ca"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.887479 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79d0-account-create-update-w29fr" event={"ID":"4136bc47-2e87-4af7-acf9-23a36c37269d","Type":"ContainerStarted","Data":"072ca5cada6215872a924ac25e120d7a4b4846fa04ede664014caaa49b27f0c3"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.890193 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf899aa6-5d99-482f-b493-56d7de87e44c" containerID="fa60c28e4cd5fd3128b80c2e4ba583fb162a3418f60a1034dc3f7be5406bd078" exitCode=0 Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.891615 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2rcd6" event={"ID":"bf899aa6-5d99-482f-b493-56d7de87e44c","Type":"ContainerDied","Data":"fa60c28e4cd5fd3128b80c2e4ba583fb162a3418f60a1034dc3f7be5406bd078"} Jan 06 14:57:09 crc kubenswrapper[4744]: I0106 14:57:09.904873 4744 scope.go:117] "RemoveContainer" containerID="6f242fe9be30286a41017c6e3c8d2750feca44df689bee4dea5898e45b5959cd" Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.044269 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f9b819c-e185-4a98-8f78-32951d751132" (UID: "7f9b819c-e185-4a98-8f78-32951d751132"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.048262 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f9b819c-e185-4a98-8f78-32951d751132" (UID: "7f9b819c-e185-4a98-8f78-32951d751132"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.053014 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.053047 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.256967 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-config" (OuterVolumeSpecName: "config") pod "7f9b819c-e185-4a98-8f78-32951d751132" (UID: "7f9b819c-e185-4a98-8f78-32951d751132"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.358951 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f9b819c-e185-4a98-8f78-32951d751132-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.515567 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-4gtkg"] Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.523560 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-4gtkg"] Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.900391 4744 generic.go:334] "Generic (PLEG): container finished" podID="218cb457-4148-4612-bbf6-f8240c136d3b" containerID="897553c86d3651136360bd9d6a4f409073678c53836195c8462b58e9ad930617" exitCode=0 Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.900472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" event={"ID":"218cb457-4148-4612-bbf6-f8240c136d3b","Type":"ContainerDied","Data":"897553c86d3651136360bd9d6a4f409073678c53836195c8462b58e9ad930617"} Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.902157 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f0f3-account-create-update-h8jq2" event={"ID":"de589768-1fff-4511-a85e-cbb4e8cb30ed","Type":"ContainerStarted","Data":"89a23289537705577d0a0e00b5060f8fc126f4bb7c59f24cf2b0282e66862b86"} Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.905247 4744 generic.go:334] "Generic (PLEG): container finished" podID="4136bc47-2e87-4af7-acf9-23a36c37269d" containerID="8a8d208339772abdef80cc2136ad05331e6fe7a98777376e84a8b2d48a796568" exitCode=0 Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.905314 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79d0-account-create-update-w29fr" event={"ID":"4136bc47-2e87-4af7-acf9-23a36c37269d","Type":"ContainerDied","Data":"8a8d208339772abdef80cc2136ad05331e6fe7a98777376e84a8b2d48a796568"} Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.906750 4744 generic.go:334] "Generic (PLEG): container finished" podID="f410252c-e5a3-4d42-ba06-febebef5a30a" containerID="ba10af281287d6f6d53e1a4f86eb77f6e5a2f7416049570f8be012f584736d87" exitCode=0 Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.906805 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bv5rj" event={"ID":"f410252c-e5a3-4d42-ba06-febebef5a30a","Type":"ContainerDied","Data":"ba10af281287d6f6d53e1a4f86eb77f6e5a2f7416049570f8be012f584736d87"} Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.908195 4744 generic.go:334] "Generic (PLEG): container finished" podID="d755b346-e89a-4ae8-a162-4fc6f8c28797" containerID="4ff9a47398dad8581d91e0b79fe09b122503f679df40098ef216cc80866153c8" exitCode=0 Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.908246 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-thfs7" event={"ID":"d755b346-e89a-4ae8-a162-4fc6f8c28797","Type":"ContainerDied","Data":"4ff9a47398dad8581d91e0b79fe09b122503f679df40098ef216cc80866153c8"} Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.909590 4744 generic.go:334] "Generic (PLEG): container finished" podID="c10fd027-3c56-4a1d-a16f-de6b641a82f4" containerID="7b1cba8e587256820a5e1e682c881c55418c80b71d42324099c2c0bafd6134d1" exitCode=0 Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.909646 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" event={"ID":"c10fd027-3c56-4a1d-a16f-de6b641a82f4","Type":"ContainerDied","Data":"7b1cba8e587256820a5e1e682c881c55418c80b71d42324099c2c0bafd6134d1"} Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.911939 4744 generic.go:334] "Generic (PLEG): container finished" podID="d7e8cff6-96c6-4fee-99ca-1f77403080b4" containerID="37171b25c1cefc9d4a58eecf515fcc2ec59aed08d23e508549e1f6bf04fdbd05" exitCode=0 Jan 06 14:57:10 crc kubenswrapper[4744]: I0106 14:57:10.912201 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8d3a-account-create-update-fhjdm" event={"ID":"d7e8cff6-96c6-4fee-99ca-1f77403080b4","Type":"ContainerDied","Data":"37171b25c1cefc9d4a58eecf515fcc2ec59aed08d23e508549e1f6bf04fdbd05"} Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.027877 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-f0f3-account-create-update-h8jq2" podStartSLOduration=10.027859236 podStartE2EDuration="10.027859236s" podCreationTimestamp="2026-01-06 14:57:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:11.024789845 +0000 UTC m=+1227.652256163" watchObservedRunningTime="2026-01-06 14:57:11.027859236 +0000 UTC m=+1227.655325554" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.499711 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.504610 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xf2lt" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.687967 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf899aa6-5d99-482f-b493-56d7de87e44c-operator-scripts\") pod \"bf899aa6-5d99-482f-b493-56d7de87e44c\" (UID: \"bf899aa6-5d99-482f-b493-56d7de87e44c\") " Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.688103 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khf8h\" (UniqueName: \"kubernetes.io/projected/ac516648-534b-4d75-ae6e-d646c0d268e8-kube-api-access-khf8h\") pod \"ac516648-534b-4d75-ae6e-d646c0d268e8\" (UID: \"ac516648-534b-4d75-ae6e-d646c0d268e8\") " Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.688134 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl2z9\" (UniqueName: \"kubernetes.io/projected/bf899aa6-5d99-482f-b493-56d7de87e44c-kube-api-access-rl2z9\") pod \"bf899aa6-5d99-482f-b493-56d7de87e44c\" (UID: \"bf899aa6-5d99-482f-b493-56d7de87e44c\") " Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.688196 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac516648-534b-4d75-ae6e-d646c0d268e8-operator-scripts\") pod \"ac516648-534b-4d75-ae6e-d646c0d268e8\" (UID: \"ac516648-534b-4d75-ae6e-d646c0d268e8\") " Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.692578 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf899aa6-5d99-482f-b493-56d7de87e44c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bf899aa6-5d99-482f-b493-56d7de87e44c" (UID: "bf899aa6-5d99-482f-b493-56d7de87e44c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.692579 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac516648-534b-4d75-ae6e-d646c0d268e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ac516648-534b-4d75-ae6e-d646c0d268e8" (UID: "ac516648-534b-4d75-ae6e-d646c0d268e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.697299 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf899aa6-5d99-482f-b493-56d7de87e44c-kube-api-access-rl2z9" (OuterVolumeSpecName: "kube-api-access-rl2z9") pod "bf899aa6-5d99-482f-b493-56d7de87e44c" (UID: "bf899aa6-5d99-482f-b493-56d7de87e44c"). InnerVolumeSpecName "kube-api-access-rl2z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.705352 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac516648-534b-4d75-ae6e-d646c0d268e8-kube-api-access-khf8h" (OuterVolumeSpecName: "kube-api-access-khf8h") pod "ac516648-534b-4d75-ae6e-d646c0d268e8" (UID: "ac516648-534b-4d75-ae6e-d646c0d268e8"). InnerVolumeSpecName "kube-api-access-khf8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.742627 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f9b819c-e185-4a98-8f78-32951d751132" path="/var/lib/kubelet/pods/7f9b819c-e185-4a98-8f78-32951d751132/volumes" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.792448 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac516648-534b-4d75-ae6e-d646c0d268e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.792477 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bf899aa6-5d99-482f-b493-56d7de87e44c-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.792486 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khf8h\" (UniqueName: \"kubernetes.io/projected/ac516648-534b-4d75-ae6e-d646c0d268e8-kube-api-access-khf8h\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.792496 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl2z9\" (UniqueName: \"kubernetes.io/projected/bf899aa6-5d99-482f-b493-56d7de87e44c-kube-api-access-rl2z9\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.921073 4744 generic.go:334] "Generic (PLEG): container finished" podID="de589768-1fff-4511-a85e-cbb4e8cb30ed" containerID="89a23289537705577d0a0e00b5060f8fc126f4bb7c59f24cf2b0282e66862b86" exitCode=0 Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.921123 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f0f3-account-create-update-h8jq2" event={"ID":"de589768-1fff-4511-a85e-cbb4e8cb30ed","Type":"ContainerDied","Data":"89a23289537705577d0a0e00b5060f8fc126f4bb7c59f24cf2b0282e66862b86"} Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.922831 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-2rcd6" event={"ID":"bf899aa6-5d99-482f-b493-56d7de87e44c","Type":"ContainerDied","Data":"367e85e3fef5ea359c3a9b5ae21d2a29735a77abfcfecce5f5849eb7d383083a"} Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.922859 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="367e85e3fef5ea359c3a9b5ae21d2a29735a77abfcfecce5f5849eb7d383083a" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.923345 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-2rcd6" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.925157 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerStarted","Data":"8106728d4589820b23f4bb38e82433fa00ceb77f299f84faa4e343985d64ece0"} Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.926910 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-xf2lt" Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.927266 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-xf2lt" event={"ID":"ac516648-534b-4d75-ae6e-d646c0d268e8","Type":"ContainerDied","Data":"ca190b206102e15d2d99e1edf1d9d8c489c28099aec63af944b3c2030f6a806e"} Jan 06 14:57:11 crc kubenswrapper[4744]: I0106 14:57:11.927289 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca190b206102e15d2d99e1edf1d9d8c489c28099aec63af944b3c2030f6a806e" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.430790 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.606990 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f410252c-e5a3-4d42-ba06-febebef5a30a-operator-scripts\") pod \"f410252c-e5a3-4d42-ba06-febebef5a30a\" (UID: \"f410252c-e5a3-4d42-ba06-febebef5a30a\") " Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.607367 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq8m4\" (UniqueName: \"kubernetes.io/projected/f410252c-e5a3-4d42-ba06-febebef5a30a-kube-api-access-sq8m4\") pod \"f410252c-e5a3-4d42-ba06-febebef5a30a\" (UID: \"f410252c-e5a3-4d42-ba06-febebef5a30a\") " Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.608302 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f410252c-e5a3-4d42-ba06-febebef5a30a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f410252c-e5a3-4d42-ba06-febebef5a30a" (UID: "f410252c-e5a3-4d42-ba06-febebef5a30a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.639791 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f410252c-e5a3-4d42-ba06-febebef5a30a-kube-api-access-sq8m4" (OuterVolumeSpecName: "kube-api-access-sq8m4") pod "f410252c-e5a3-4d42-ba06-febebef5a30a" (UID: "f410252c-e5a3-4d42-ba06-febebef5a30a"). InnerVolumeSpecName "kube-api-access-sq8m4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.709367 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq8m4\" (UniqueName: \"kubernetes.io/projected/f410252c-e5a3-4d42-ba06-febebef5a30a-kube-api-access-sq8m4\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.709405 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f410252c-e5a3-4d42-ba06-febebef5a30a-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.939414 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bv5rj" event={"ID":"f410252c-e5a3-4d42-ba06-febebef5a30a","Type":"ContainerDied","Data":"e1f1bf5fdd3ae4f8f066081bc6550d7988bf58f23f9a0163010cd549c7b70cdd"} Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.939472 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1f1bf5fdd3ae4f8f066081bc6550d7988bf58f23f9a0163010cd549c7b70cdd" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.939434 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bv5rj" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.941092 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-r4r76" podUID="7f86f187-37db-488e-b6f6-22a5becacecd" containerName="ovn-controller" probeResult="failure" output=< Jan 06 14:57:12 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 06 14:57:12 crc kubenswrapper[4744]: > Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.941616 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-thfs7" event={"ID":"d755b346-e89a-4ae8-a162-4fc6f8c28797","Type":"ContainerDied","Data":"8c500d01712b6c097d0eab8a9c6f3e261342b7437faf36f2da421338175701a8"} Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.941665 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c500d01712b6c097d0eab8a9c6f3e261342b7437faf36f2da421338175701a8" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.944349 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" event={"ID":"c10fd027-3c56-4a1d-a16f-de6b641a82f4","Type":"ContainerDied","Data":"17137168f6ccfc7e674b4bccba59ed5e27bb7941d029088f9ceaa674b5014cdd"} Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.944375 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17137168f6ccfc7e674b4bccba59ed5e27bb7941d029088f9ceaa674b5014cdd" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.947486 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8d3a-account-create-update-fhjdm" event={"ID":"d7e8cff6-96c6-4fee-99ca-1f77403080b4","Type":"ContainerDied","Data":"096cdbf6781798dc9ba817cf19711b776802cde071ad51c5c9bce9e21237ca71"} Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.947508 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="096cdbf6781798dc9ba817cf19711b776802cde071ad51c5c9bce9e21237ca71" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.953283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" event={"ID":"218cb457-4148-4612-bbf6-f8240c136d3b","Type":"ContainerDied","Data":"dfc823a84423cdf91639caed00468369ea80f1a2deb68af5b189b685d2ad57ca"} Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.953341 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfc823a84423cdf91639caed00468369ea80f1a2deb68af5b189b685d2ad57ca" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.955472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-79d0-account-create-update-w29fr" event={"ID":"4136bc47-2e87-4af7-acf9-23a36c37269d","Type":"ContainerDied","Data":"072ca5cada6215872a924ac25e120d7a4b4846fa04ede664014caaa49b27f0c3"} Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.955512 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="072ca5cada6215872a924ac25e120d7a4b4846fa04ede664014caaa49b27f0c3" Jan 06 14:57:12 crc kubenswrapper[4744]: I0106 14:57:12.989540 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.005315 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.012791 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.026091 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-thfs7" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.034592 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.119122 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8cff6-96c6-4fee-99ca-1f77403080b4-operator-scripts\") pod \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\" (UID: \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.119246 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d755b346-e89a-4ae8-a162-4fc6f8c28797-operator-scripts\") pod \"d755b346-e89a-4ae8-a162-4fc6f8c28797\" (UID: \"d755b346-e89a-4ae8-a162-4fc6f8c28797\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.119271 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqbfs\" (UniqueName: \"kubernetes.io/projected/c10fd027-3c56-4a1d-a16f-de6b641a82f4-kube-api-access-vqbfs\") pod \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\" (UID: \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.119333 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdrpq\" (UniqueName: \"kubernetes.io/projected/d755b346-e89a-4ae8-a162-4fc6f8c28797-kube-api-access-sdrpq\") pod \"d755b346-e89a-4ae8-a162-4fc6f8c28797\" (UID: \"d755b346-e89a-4ae8-a162-4fc6f8c28797\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.119355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzrf9\" (UniqueName: \"kubernetes.io/projected/d7e8cff6-96c6-4fee-99ca-1f77403080b4-kube-api-access-kzrf9\") pod \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\" (UID: \"d7e8cff6-96c6-4fee-99ca-1f77403080b4\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.119426 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99zhz\" (UniqueName: \"kubernetes.io/projected/218cb457-4148-4612-bbf6-f8240c136d3b-kube-api-access-99zhz\") pod \"218cb457-4148-4612-bbf6-f8240c136d3b\" (UID: \"218cb457-4148-4612-bbf6-f8240c136d3b\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.119547 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c10fd027-3c56-4a1d-a16f-de6b641a82f4-operator-scripts\") pod \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\" (UID: \"c10fd027-3c56-4a1d-a16f-de6b641a82f4\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.119582 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218cb457-4148-4612-bbf6-f8240c136d3b-operator-scripts\") pod \"218cb457-4148-4612-bbf6-f8240c136d3b\" (UID: \"218cb457-4148-4612-bbf6-f8240c136d3b\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.119675 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e8cff6-96c6-4fee-99ca-1f77403080b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7e8cff6-96c6-4fee-99ca-1f77403080b4" (UID: "d7e8cff6-96c6-4fee-99ca-1f77403080b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.120341 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/218cb457-4148-4612-bbf6-f8240c136d3b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "218cb457-4148-4612-bbf6-f8240c136d3b" (UID: "218cb457-4148-4612-bbf6-f8240c136d3b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.120348 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c10fd027-3c56-4a1d-a16f-de6b641a82f4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c10fd027-3c56-4a1d-a16f-de6b641a82f4" (UID: "c10fd027-3c56-4a1d-a16f-de6b641a82f4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.121411 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7e8cff6-96c6-4fee-99ca-1f77403080b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.121438 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c10fd027-3c56-4a1d-a16f-de6b641a82f4-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.121449 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/218cb457-4148-4612-bbf6-f8240c136d3b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.121845 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d755b346-e89a-4ae8-a162-4fc6f8c28797-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d755b346-e89a-4ae8-a162-4fc6f8c28797" (UID: "d755b346-e89a-4ae8-a162-4fc6f8c28797"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.125137 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d755b346-e89a-4ae8-a162-4fc6f8c28797-kube-api-access-sdrpq" (OuterVolumeSpecName: "kube-api-access-sdrpq") pod "d755b346-e89a-4ae8-a162-4fc6f8c28797" (UID: "d755b346-e89a-4ae8-a162-4fc6f8c28797"). InnerVolumeSpecName "kube-api-access-sdrpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.127356 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c10fd027-3c56-4a1d-a16f-de6b641a82f4-kube-api-access-vqbfs" (OuterVolumeSpecName: "kube-api-access-vqbfs") pod "c10fd027-3c56-4a1d-a16f-de6b641a82f4" (UID: "c10fd027-3c56-4a1d-a16f-de6b641a82f4"). InnerVolumeSpecName "kube-api-access-vqbfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.127679 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e8cff6-96c6-4fee-99ca-1f77403080b4-kube-api-access-kzrf9" (OuterVolumeSpecName: "kube-api-access-kzrf9") pod "d7e8cff6-96c6-4fee-99ca-1f77403080b4" (UID: "d7e8cff6-96c6-4fee-99ca-1f77403080b4"). InnerVolumeSpecName "kube-api-access-kzrf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.127891 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/218cb457-4148-4612-bbf6-f8240c136d3b-kube-api-access-99zhz" (OuterVolumeSpecName: "kube-api-access-99zhz") pod "218cb457-4148-4612-bbf6-f8240c136d3b" (UID: "218cb457-4148-4612-bbf6-f8240c136d3b"). InnerVolumeSpecName "kube-api-access-99zhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.222878 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4136bc47-2e87-4af7-acf9-23a36c37269d-operator-scripts\") pod \"4136bc47-2e87-4af7-acf9-23a36c37269d\" (UID: \"4136bc47-2e87-4af7-acf9-23a36c37269d\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.222965 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6cwd\" (UniqueName: \"kubernetes.io/projected/4136bc47-2e87-4af7-acf9-23a36c37269d-kube-api-access-w6cwd\") pod \"4136bc47-2e87-4af7-acf9-23a36c37269d\" (UID: \"4136bc47-2e87-4af7-acf9-23a36c37269d\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.223573 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4136bc47-2e87-4af7-acf9-23a36c37269d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4136bc47-2e87-4af7-acf9-23a36c37269d" (UID: "4136bc47-2e87-4af7-acf9-23a36c37269d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.223648 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d755b346-e89a-4ae8-a162-4fc6f8c28797-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.223672 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqbfs\" (UniqueName: \"kubernetes.io/projected/c10fd027-3c56-4a1d-a16f-de6b641a82f4-kube-api-access-vqbfs\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.223690 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdrpq\" (UniqueName: \"kubernetes.io/projected/d755b346-e89a-4ae8-a162-4fc6f8c28797-kube-api-access-sdrpq\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.223700 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzrf9\" (UniqueName: \"kubernetes.io/projected/d7e8cff6-96c6-4fee-99ca-1f77403080b4-kube-api-access-kzrf9\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.223712 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99zhz\" (UniqueName: \"kubernetes.io/projected/218cb457-4148-4612-bbf6-f8240c136d3b-kube-api-access-99zhz\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.227965 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4136bc47-2e87-4af7-acf9-23a36c37269d-kube-api-access-w6cwd" (OuterVolumeSpecName: "kube-api-access-w6cwd") pod "4136bc47-2e87-4af7-acf9-23a36c37269d" (UID: "4136bc47-2e87-4af7-acf9-23a36c37269d"). InnerVolumeSpecName "kube-api-access-w6cwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.325618 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4136bc47-2e87-4af7-acf9-23a36c37269d-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.325651 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6cwd\" (UniqueName: \"kubernetes.io/projected/4136bc47-2e87-4af7-acf9-23a36c37269d-kube-api-access-w6cwd\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.353097 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.530630 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nczl5\" (UniqueName: \"kubernetes.io/projected/de589768-1fff-4511-a85e-cbb4e8cb30ed-kube-api-access-nczl5\") pod \"de589768-1fff-4511-a85e-cbb4e8cb30ed\" (UID: \"de589768-1fff-4511-a85e-cbb4e8cb30ed\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.531037 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de589768-1fff-4511-a85e-cbb4e8cb30ed-operator-scripts\") pod \"de589768-1fff-4511-a85e-cbb4e8cb30ed\" (UID: \"de589768-1fff-4511-a85e-cbb4e8cb30ed\") " Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.531548 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de589768-1fff-4511-a85e-cbb4e8cb30ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "de589768-1fff-4511-a85e-cbb4e8cb30ed" (UID: "de589768-1fff-4511-a85e-cbb4e8cb30ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.531720 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de589768-1fff-4511-a85e-cbb4e8cb30ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.534397 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de589768-1fff-4511-a85e-cbb4e8cb30ed-kube-api-access-nczl5" (OuterVolumeSpecName: "kube-api-access-nczl5") pod "de589768-1fff-4511-a85e-cbb4e8cb30ed" (UID: "de589768-1fff-4511-a85e-cbb4e8cb30ed"). InnerVolumeSpecName "kube-api-access-nczl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.633413 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nczl5\" (UniqueName: \"kubernetes.io/projected/de589768-1fff-4511-a85e-cbb4e8cb30ed-kube-api-access-nczl5\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.984843 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f0f3-account-create-update-h8jq2" event={"ID":"de589768-1fff-4511-a85e-cbb4e8cb30ed","Type":"ContainerDied","Data":"3b38be9543a2e95b2af32a5b267ab930ed4bab03725e3a733fae1a58beb14d17"} Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.985058 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b38be9543a2e95b2af32a5b267ab930ed4bab03725e3a733fae1a58beb14d17" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.985207 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f0f3-account-create-update-h8jq2" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.993968 4744 generic.go:334] "Generic (PLEG): container finished" podID="82b9e690-d1d7-473d-99d1-20be1130dfd3" containerID="1798a88497d35559965f70f8539e90b4e5019f317139ea4a0a68b98930885e8f" exitCode=0 Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.994138 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-79d0-account-create-update-w29fr" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.994357 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-x5r7t" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.995373 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8d3a-account-create-update-fhjdm" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.995807 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-thfs7" Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.996380 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fq864" event={"ID":"82b9e690-d1d7-473d-99d1-20be1130dfd3","Type":"ContainerDied","Data":"1798a88497d35559965f70f8539e90b4e5019f317139ea4a0a68b98930885e8f"} Jan 06 14:57:13 crc kubenswrapper[4744]: I0106 14:57:13.996499 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-6c8a-account-create-update-7b4pf" Jan 06 14:57:14 crc kubenswrapper[4744]: I0106 14:57:14.423447 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:57:14 crc kubenswrapper[4744]: I0106 14:57:14.423762 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:57:14 crc kubenswrapper[4744]: I0106 14:57:14.628913 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-xf2lt"] Jan 06 14:57:14 crc kubenswrapper[4744]: I0106 14:57:14.638130 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-xf2lt"] Jan 06 14:57:15 crc kubenswrapper[4744]: I0106 14:57:15.803730 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac516648-534b-4d75-ae6e-d646c0d268e8" path="/var/lib/kubelet/pods/ac516648-534b-4d75-ae6e-d646c0d268e8/volumes" Jan 06 14:57:15 crc kubenswrapper[4744]: I0106 14:57:15.975031 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.014091 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-fq864" event={"ID":"82b9e690-d1d7-473d-99d1-20be1130dfd3","Type":"ContainerDied","Data":"ebb9c082be7a14f3d4235dfc16de25ee19935cf56bba639b9b0efbc811a29b35"} Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.014139 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebb9c082be7a14f3d4235dfc16de25ee19935cf56bba639b9b0efbc811a29b35" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.014260 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-fq864" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.083317 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/82b9e690-d1d7-473d-99d1-20be1130dfd3-etc-swift\") pod \"82b9e690-d1d7-473d-99d1-20be1130dfd3\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.083895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-dispersionconf\") pod \"82b9e690-d1d7-473d-99d1-20be1130dfd3\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.084498 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82b9e690-d1d7-473d-99d1-20be1130dfd3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "82b9e690-d1d7-473d-99d1-20be1130dfd3" (UID: "82b9e690-d1d7-473d-99d1-20be1130dfd3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.088601 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg6cf\" (UniqueName: \"kubernetes.io/projected/82b9e690-d1d7-473d-99d1-20be1130dfd3-kube-api-access-qg6cf\") pod \"82b9e690-d1d7-473d-99d1-20be1130dfd3\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.088768 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-scripts\") pod \"82b9e690-d1d7-473d-99d1-20be1130dfd3\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.088880 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-combined-ca-bundle\") pod \"82b9e690-d1d7-473d-99d1-20be1130dfd3\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.089016 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-swiftconf\") pod \"82b9e690-d1d7-473d-99d1-20be1130dfd3\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.089086 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-ring-data-devices\") pod \"82b9e690-d1d7-473d-99d1-20be1130dfd3\" (UID: \"82b9e690-d1d7-473d-99d1-20be1130dfd3\") " Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.090353 4744 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/82b9e690-d1d7-473d-99d1-20be1130dfd3-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.091556 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "82b9e690-d1d7-473d-99d1-20be1130dfd3" (UID: "82b9e690-d1d7-473d-99d1-20be1130dfd3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.097453 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82b9e690-d1d7-473d-99d1-20be1130dfd3-kube-api-access-qg6cf" (OuterVolumeSpecName: "kube-api-access-qg6cf") pod "82b9e690-d1d7-473d-99d1-20be1130dfd3" (UID: "82b9e690-d1d7-473d-99d1-20be1130dfd3"). InnerVolumeSpecName "kube-api-access-qg6cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.101903 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "82b9e690-d1d7-473d-99d1-20be1130dfd3" (UID: "82b9e690-d1d7-473d-99d1-20be1130dfd3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.128476 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82b9e690-d1d7-473d-99d1-20be1130dfd3" (UID: "82b9e690-d1d7-473d-99d1-20be1130dfd3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.131126 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "82b9e690-d1d7-473d-99d1-20be1130dfd3" (UID: "82b9e690-d1d7-473d-99d1-20be1130dfd3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.140817 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-scripts" (OuterVolumeSpecName: "scripts") pod "82b9e690-d1d7-473d-99d1-20be1130dfd3" (UID: "82b9e690-d1d7-473d-99d1-20be1130dfd3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.192830 4744 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.192886 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg6cf\" (UniqueName: \"kubernetes.io/projected/82b9e690-d1d7-473d-99d1-20be1130dfd3-kube-api-access-qg6cf\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.192912 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.192931 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.192951 4744 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/82b9e690-d1d7-473d-99d1-20be1130dfd3-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.192968 4744 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/82b9e690-d1d7-473d-99d1-20be1130dfd3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.499241 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.504268 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c603cde3-b7b0-4851-99df-2abb0aa6b1c4-etc-swift\") pod \"swift-storage-0\" (UID: \"c603cde3-b7b0-4851-99df-2abb0aa6b1c4\") " pod="openstack/swift-storage-0" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.747573 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.894363 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-68rr4"] Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.894855 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218cb457-4148-4612-bbf6-f8240c136d3b" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.894870 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="218cb457-4148-4612-bbf6-f8240c136d3b" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.894892 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac516648-534b-4d75-ae6e-d646c0d268e8" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.894901 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac516648-534b-4d75-ae6e-d646c0d268e8" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.894911 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf899aa6-5d99-482f-b493-56d7de87e44c" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.894944 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf899aa6-5d99-482f-b493-56d7de87e44c" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.894954 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de589768-1fff-4511-a85e-cbb4e8cb30ed" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.894962 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="de589768-1fff-4511-a85e-cbb4e8cb30ed" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.894979 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e8cff6-96c6-4fee-99ca-1f77403080b4" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.894988 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e8cff6-96c6-4fee-99ca-1f77403080b4" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.895005 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f9b819c-e185-4a98-8f78-32951d751132" containerName="init" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895014 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f9b819c-e185-4a98-8f78-32951d751132" containerName="init" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.895029 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f9b819c-e185-4a98-8f78-32951d751132" containerName="dnsmasq-dns" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895037 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f9b819c-e185-4a98-8f78-32951d751132" containerName="dnsmasq-dns" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.895049 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d755b346-e89a-4ae8-a162-4fc6f8c28797" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895067 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d755b346-e89a-4ae8-a162-4fc6f8c28797" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.895083 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f410252c-e5a3-4d42-ba06-febebef5a30a" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895091 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f410252c-e5a3-4d42-ba06-febebef5a30a" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.895104 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b9e690-d1d7-473d-99d1-20be1130dfd3" containerName="swift-ring-rebalance" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895112 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b9e690-d1d7-473d-99d1-20be1130dfd3" containerName="swift-ring-rebalance" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.895132 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4136bc47-2e87-4af7-acf9-23a36c37269d" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895140 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4136bc47-2e87-4af7-acf9-23a36c37269d" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: E0106 14:57:16.895156 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c10fd027-3c56-4a1d-a16f-de6b641a82f4" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895178 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c10fd027-3c56-4a1d-a16f-de6b641a82f4" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895396 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf899aa6-5d99-482f-b493-56d7de87e44c" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895408 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac516648-534b-4d75-ae6e-d646c0d268e8" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895426 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="218cb457-4148-4612-bbf6-f8240c136d3b" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895438 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f9b819c-e185-4a98-8f78-32951d751132" containerName="dnsmasq-dns" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895450 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f410252c-e5a3-4d42-ba06-febebef5a30a" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895459 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="de589768-1fff-4511-a85e-cbb4e8cb30ed" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895473 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="82b9e690-d1d7-473d-99d1-20be1130dfd3" containerName="swift-ring-rebalance" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895486 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4136bc47-2e87-4af7-acf9-23a36c37269d" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895499 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e8cff6-96c6-4fee-99ca-1f77403080b4" containerName="mariadb-account-create-update" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895518 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d755b346-e89a-4ae8-a162-4fc6f8c28797" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.895535 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c10fd027-3c56-4a1d-a16f-de6b641a82f4" containerName="mariadb-database-create" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.896414 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.898725 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.905194 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lxm4v" Jan 06 14:57:16 crc kubenswrapper[4744]: I0106 14:57:16.913904 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-68rr4"] Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.011270 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hstt2\" (UniqueName: \"kubernetes.io/projected/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-kube-api-access-hstt2\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.011437 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-db-sync-config-data\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.011561 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-config-data\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.011622 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-combined-ca-bundle\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.041896 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerStarted","Data":"a268c49ab4e371be422a1a35ca45cacd6c76688539be436d8fc478a71d2f361d"} Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.074747 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=21.543472445 podStartE2EDuration="1m14.074730572s" podCreationTimestamp="2026-01-06 14:56:03 +0000 UTC" firstStartedPulling="2026-01-06 14:56:24.00478761 +0000 UTC m=+1180.632253928" lastFinishedPulling="2026-01-06 14:57:16.536045717 +0000 UTC m=+1233.163512055" observedRunningTime="2026-01-06 14:57:17.074176328 +0000 UTC m=+1233.701642646" watchObservedRunningTime="2026-01-06 14:57:17.074730572 +0000 UTC m=+1233.702196890" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.114662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-config-data\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.114876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-combined-ca-bundle\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.114979 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hstt2\" (UniqueName: \"kubernetes.io/projected/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-kube-api-access-hstt2\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.115113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-db-sync-config-data\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.125792 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-config-data\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.126355 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-db-sync-config-data\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.126439 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-combined-ca-bundle\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.136798 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hstt2\" (UniqueName: \"kubernetes.io/projected/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-kube-api-access-hstt2\") pod \"glance-db-sync-68rr4\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.216799 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-68rr4" Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.458671 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.871882 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-68rr4"] Jan 06 14:57:17 crc kubenswrapper[4744]: I0106 14:57:17.982384 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-r4r76" podUID="7f86f187-37db-488e-b6f6-22a5becacecd" containerName="ovn-controller" probeResult="failure" output=< Jan 06 14:57:17 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 06 14:57:17 crc kubenswrapper[4744]: > Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.051367 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-68rr4" event={"ID":"cb6ad61a-8190-4d4d-987c-f609c1e8cf44","Type":"ContainerStarted","Data":"3dd926f3cf954db0de06835d3632a80433b3a1b42f8f4c0ca87fcf7508d93ef3"} Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.052916 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"2c66ad65c2c268d437ff521c4a6c5a7357df0e95efef1a50b9bf2c951a4a1a04"} Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.394784 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.128:5671: connect: connection refused" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.723799 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.732647 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.748778 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf"] Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.750060 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.756742 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf"] Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.846257 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67d656c5-0291-4edd-b1cd-304b1bdce02b-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-sfjpf\" (UID: \"67d656c5-0291-4edd-b1cd-304b1bdce02b\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.846428 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxzrh\" (UniqueName: \"kubernetes.io/projected/67d656c5-0291-4edd-b1cd-304b1bdce02b-kube-api-access-wxzrh\") pod \"mysqld-exporter-openstack-cell1-db-create-sfjpf\" (UID: \"67d656c5-0291-4edd-b1cd-304b1bdce02b\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.949006 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67d656c5-0291-4edd-b1cd-304b1bdce02b-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-sfjpf\" (UID: \"67d656c5-0291-4edd-b1cd-304b1bdce02b\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.949236 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxzrh\" (UniqueName: \"kubernetes.io/projected/67d656c5-0291-4edd-b1cd-304b1bdce02b-kube-api-access-wxzrh\") pod \"mysqld-exporter-openstack-cell1-db-create-sfjpf\" (UID: \"67d656c5-0291-4edd-b1cd-304b1bdce02b\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.949857 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67d656c5-0291-4edd-b1cd-304b1bdce02b-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-sfjpf\" (UID: \"67d656c5-0291-4edd-b1cd-304b1bdce02b\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.963045 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-08dd-account-create-update-hph9l"] Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.965976 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.972829 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.981992 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-08dd-account-create-update-hph9l"] Jan 06 14:57:18 crc kubenswrapper[4744]: I0106 14:57:18.988800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxzrh\" (UniqueName: \"kubernetes.io/projected/67d656c5-0291-4edd-b1cd-304b1bdce02b-kube-api-access-wxzrh\") pod \"mysqld-exporter-openstack-cell1-db-create-sfjpf\" (UID: \"67d656c5-0291-4edd-b1cd-304b1bdce02b\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.051400 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.051527 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlw9f\" (UniqueName: \"kubernetes.io/projected/f59d0597-674a-4db8-b70c-1307186cf198-kube-api-access-rlw9f\") pod \"mysqld-exporter-08dd-account-create-update-hph9l\" (UID: \"f59d0597-674a-4db8-b70c-1307186cf198\") " pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.051613 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f59d0597-674a-4db8-b70c-1307186cf198-operator-scripts\") pod \"mysqld-exporter-08dd-account-create-update-hph9l\" (UID: \"f59d0597-674a-4db8-b70c-1307186cf198\") " pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.072805 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.153385 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlw9f\" (UniqueName: \"kubernetes.io/projected/f59d0597-674a-4db8-b70c-1307186cf198-kube-api-access-rlw9f\") pod \"mysqld-exporter-08dd-account-create-update-hph9l\" (UID: \"f59d0597-674a-4db8-b70c-1307186cf198\") " pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.153431 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f59d0597-674a-4db8-b70c-1307186cf198-operator-scripts\") pod \"mysqld-exporter-08dd-account-create-update-hph9l\" (UID: \"f59d0597-674a-4db8-b70c-1307186cf198\") " pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.154618 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f59d0597-674a-4db8-b70c-1307186cf198-operator-scripts\") pod \"mysqld-exporter-08dd-account-create-update-hph9l\" (UID: \"f59d0597-674a-4db8-b70c-1307186cf198\") " pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.186218 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlw9f\" (UniqueName: \"kubernetes.io/projected/f59d0597-674a-4db8-b70c-1307186cf198-kube-api-access-rlw9f\") pod \"mysqld-exporter-08dd-account-create-update-hph9l\" (UID: \"f59d0597-674a-4db8-b70c-1307186cf198\") " pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.306950 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.644284 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-8lng4"] Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.646232 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.649552 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.659149 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-8lng4"] Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.677772 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d7e690-64ff-4302-ab9d-57685f2abc88-operator-scripts\") pod \"root-account-create-update-8lng4\" (UID: \"b5d7e690-64ff-4302-ab9d-57685f2abc88\") " pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.677834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxqb7\" (UniqueName: \"kubernetes.io/projected/b5d7e690-64ff-4302-ab9d-57685f2abc88-kube-api-access-gxqb7\") pod \"root-account-create-update-8lng4\" (UID: \"b5d7e690-64ff-4302-ab9d-57685f2abc88\") " pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.699025 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf"] Jan 06 14:57:19 crc kubenswrapper[4744]: W0106 14:57:19.699234 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67d656c5_0291_4edd_b1cd_304b1bdce02b.slice/crio-c1b7b6d836d94ca5f7963389ffd5107ff6d62b92efdb2a069d5f16bd9bfc3ab3 WatchSource:0}: Error finding container c1b7b6d836d94ca5f7963389ffd5107ff6d62b92efdb2a069d5f16bd9bfc3ab3: Status 404 returned error can't find the container with id c1b7b6d836d94ca5f7963389ffd5107ff6d62b92efdb2a069d5f16bd9bfc3ab3 Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.782152 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxqb7\" (UniqueName: \"kubernetes.io/projected/b5d7e690-64ff-4302-ab9d-57685f2abc88-kube-api-access-gxqb7\") pod \"root-account-create-update-8lng4\" (UID: \"b5d7e690-64ff-4302-ab9d-57685f2abc88\") " pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.783616 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d7e690-64ff-4302-ab9d-57685f2abc88-operator-scripts\") pod \"root-account-create-update-8lng4\" (UID: \"b5d7e690-64ff-4302-ab9d-57685f2abc88\") " pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.785518 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d7e690-64ff-4302-ab9d-57685f2abc88-operator-scripts\") pod \"root-account-create-update-8lng4\" (UID: \"b5d7e690-64ff-4302-ab9d-57685f2abc88\") " pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.807624 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxqb7\" (UniqueName: \"kubernetes.io/projected/b5d7e690-64ff-4302-ab9d-57685f2abc88-kube-api-access-gxqb7\") pod \"root-account-create-update-8lng4\" (UID: \"b5d7e690-64ff-4302-ab9d-57685f2abc88\") " pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.819618 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.819722 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.824970 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-08dd-account-create-update-hph9l"] Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.836677 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:19 crc kubenswrapper[4744]: W0106 14:57:19.850727 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf59d0597_674a_4db8_b70c_1307186cf198.slice/crio-6f3aa15ea685ec06c61ed0cf793177d9fd54570da5e2385bb6a027ac3de18473 WatchSource:0}: Error finding container 6f3aa15ea685ec06c61ed0cf793177d9fd54570da5e2385bb6a027ac3de18473: Status 404 returned error can't find the container with id 6f3aa15ea685ec06c61ed0cf793177d9fd54570da5e2385bb6a027ac3de18473 Jan 06 14:57:19 crc kubenswrapper[4744]: I0106 14:57:19.967872 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:20 crc kubenswrapper[4744]: I0106 14:57:20.088217 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" event={"ID":"f59d0597-674a-4db8-b70c-1307186cf198","Type":"ContainerStarted","Data":"6f3aa15ea685ec06c61ed0cf793177d9fd54570da5e2385bb6a027ac3de18473"} Jan 06 14:57:20 crc kubenswrapper[4744]: I0106 14:57:20.100557 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"ff1c01f03033787db48728070bfdc0ae81c588b5595806e78cb1bc366d0d42f2"} Jan 06 14:57:20 crc kubenswrapper[4744]: I0106 14:57:20.100604 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"ee7ee010dabdea3851338718c2df3d60efa0edb9aa53418121ac61ed28f735c1"} Jan 06 14:57:20 crc kubenswrapper[4744]: I0106 14:57:20.104385 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" event={"ID":"67d656c5-0291-4edd-b1cd-304b1bdce02b","Type":"ContainerStarted","Data":"ceb63e211536bb28cc81a6b2ffd399f52ebccc4932b59313e1a048925e42974d"} Jan 06 14:57:20 crc kubenswrapper[4744]: I0106 14:57:20.104422 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" event={"ID":"67d656c5-0291-4edd-b1cd-304b1bdce02b","Type":"ContainerStarted","Data":"c1b7b6d836d94ca5f7963389ffd5107ff6d62b92efdb2a069d5f16bd9bfc3ab3"} Jan 06 14:57:20 crc kubenswrapper[4744]: I0106 14:57:20.106669 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:20 crc kubenswrapper[4744]: I0106 14:57:20.131040 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" podStartSLOduration=2.131019845 podStartE2EDuration="2.131019845s" podCreationTimestamp="2026-01-06 14:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:20.124672877 +0000 UTC m=+1236.752139195" watchObservedRunningTime="2026-01-06 14:57:20.131019845 +0000 UTC m=+1236.758486163" Jan 06 14:57:20 crc kubenswrapper[4744]: I0106 14:57:20.468837 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-8lng4"] Jan 06 14:57:20 crc kubenswrapper[4744]: W0106 14:57:20.478866 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5d7e690_64ff_4302_ab9d_57685f2abc88.slice/crio-c5a83067aacd703aa19c2b2b6efeb80b0fcd5dc4d971fba9d541bb18a5c21f06 WatchSource:0}: Error finding container c5a83067aacd703aa19c2b2b6efeb80b0fcd5dc4d971fba9d541bb18a5c21f06: Status 404 returned error can't find the container with id c5a83067aacd703aa19c2b2b6efeb80b0fcd5dc4d971fba9d541bb18a5c21f06 Jan 06 14:57:21 crc kubenswrapper[4744]: I0106 14:57:21.117799 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8lng4" event={"ID":"b5d7e690-64ff-4302-ab9d-57685f2abc88","Type":"ContainerStarted","Data":"4f27466bb5b9291a0e7e28e0c2cfdf7c509982e49faba5b0ded99d85d4e719fc"} Jan 06 14:57:21 crc kubenswrapper[4744]: I0106 14:57:21.118030 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8lng4" event={"ID":"b5d7e690-64ff-4302-ab9d-57685f2abc88","Type":"ContainerStarted","Data":"c5a83067aacd703aa19c2b2b6efeb80b0fcd5dc4d971fba9d541bb18a5c21f06"} Jan 06 14:57:21 crc kubenswrapper[4744]: I0106 14:57:21.122729 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" event={"ID":"f59d0597-674a-4db8-b70c-1307186cf198","Type":"ContainerStarted","Data":"f70b7515706c1872c6ba3959ae5e440e2879d46ac115e47926c935ae30fd139e"} Jan 06 14:57:21 crc kubenswrapper[4744]: I0106 14:57:21.128339 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"9afc4de905b7c535a4dbf807a6a35dd5674b02b0989fc8c33ea6fd5ad92823b9"} Jan 06 14:57:21 crc kubenswrapper[4744]: I0106 14:57:21.128366 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"b2d4ca5c978bbb5ea5f9ee5a30961ddb0ebbcba48fadfc3aa74b0aa76f91df2a"} Jan 06 14:57:21 crc kubenswrapper[4744]: I0106 14:57:21.132590 4744 generic.go:334] "Generic (PLEG): container finished" podID="67d656c5-0291-4edd-b1cd-304b1bdce02b" containerID="ceb63e211536bb28cc81a6b2ffd399f52ebccc4932b59313e1a048925e42974d" exitCode=0 Jan 06 14:57:21 crc kubenswrapper[4744]: I0106 14:57:21.132741 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" event={"ID":"67d656c5-0291-4edd-b1cd-304b1bdce02b","Type":"ContainerDied","Data":"ceb63e211536bb28cc81a6b2ffd399f52ebccc4932b59313e1a048925e42974d"} Jan 06 14:57:21 crc kubenswrapper[4744]: I0106 14:57:21.140485 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-8lng4" podStartSLOduration=2.140469659 podStartE2EDuration="2.140469659s" podCreationTimestamp="2026-01-06 14:57:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:21.13749025 +0000 UTC m=+1237.764956568" watchObservedRunningTime="2026-01-06 14:57:21.140469659 +0000 UTC m=+1237.767935977" Jan 06 14:57:21 crc kubenswrapper[4744]: I0106 14:57:21.175642 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" podStartSLOduration=3.175622028 podStartE2EDuration="3.175622028s" podCreationTimestamp="2026-01-06 14:57:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:21.169544068 +0000 UTC m=+1237.797010386" watchObservedRunningTime="2026-01-06 14:57:21.175622028 +0000 UTC m=+1237.803088346" Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.576217 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.645267 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxzrh\" (UniqueName: \"kubernetes.io/projected/67d656c5-0291-4edd-b1cd-304b1bdce02b-kube-api-access-wxzrh\") pod \"67d656c5-0291-4edd-b1cd-304b1bdce02b\" (UID: \"67d656c5-0291-4edd-b1cd-304b1bdce02b\") " Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.645848 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67d656c5-0291-4edd-b1cd-304b1bdce02b-operator-scripts\") pod \"67d656c5-0291-4edd-b1cd-304b1bdce02b\" (UID: \"67d656c5-0291-4edd-b1cd-304b1bdce02b\") " Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.646766 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d656c5-0291-4edd-b1cd-304b1bdce02b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "67d656c5-0291-4edd-b1cd-304b1bdce02b" (UID: "67d656c5-0291-4edd-b1cd-304b1bdce02b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.669088 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d656c5-0291-4edd-b1cd-304b1bdce02b-kube-api-access-wxzrh" (OuterVolumeSpecName: "kube-api-access-wxzrh") pod "67d656c5-0291-4edd-b1cd-304b1bdce02b" (UID: "67d656c5-0291-4edd-b1cd-304b1bdce02b"). InnerVolumeSpecName "kube-api-access-wxzrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.747725 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxzrh\" (UniqueName: \"kubernetes.io/projected/67d656c5-0291-4edd-b1cd-304b1bdce02b-kube-api-access-wxzrh\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.747756 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/67d656c5-0291-4edd-b1cd-304b1bdce02b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.884042 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-r4r76" podUID="7f86f187-37db-488e-b6f6-22a5becacecd" containerName="ovn-controller" probeResult="failure" output=< Jan 06 14:57:22 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 06 14:57:22 crc kubenswrapper[4744]: > Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.916920 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:57:22 crc kubenswrapper[4744]: I0106 14:57:22.945663 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-qcph6" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.172917 4744 generic.go:334] "Generic (PLEG): container finished" podID="f59d0597-674a-4db8-b70c-1307186cf198" containerID="f70b7515706c1872c6ba3959ae5e440e2879d46ac115e47926c935ae30fd139e" exitCode=0 Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.172995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" event={"ID":"f59d0597-674a-4db8-b70c-1307186cf198","Type":"ContainerDied","Data":"f70b7515706c1872c6ba3959ae5e440e2879d46ac115e47926c935ae30fd139e"} Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.178725 4744 generic.go:334] "Generic (PLEG): container finished" podID="b5d7e690-64ff-4302-ab9d-57685f2abc88" containerID="4f27466bb5b9291a0e7e28e0c2cfdf7c509982e49faba5b0ded99d85d4e719fc" exitCode=0 Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.178808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8lng4" event={"ID":"b5d7e690-64ff-4302-ab9d-57685f2abc88","Type":"ContainerDied","Data":"4f27466bb5b9291a0e7e28e0c2cfdf7c509982e49faba5b0ded99d85d4e719fc"} Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.185126 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.186248 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf" event={"ID":"67d656c5-0291-4edd-b1cd-304b1bdce02b","Type":"ContainerDied","Data":"c1b7b6d836d94ca5f7963389ffd5107ff6d62b92efdb2a069d5f16bd9bfc3ab3"} Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.186312 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1b7b6d836d94ca5f7963389ffd5107ff6d62b92efdb2a069d5f16bd9bfc3ab3" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.240813 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.241083 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="prometheus" containerID="cri-o://e8690c660760c008bf10dfaba0798d89793d116fd6a52d58128117656d2a976b" gracePeriod=600 Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.241357 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="thanos-sidecar" containerID="cri-o://a268c49ab4e371be422a1a35ca45cacd6c76688539be436d8fc478a71d2f361d" gracePeriod=600 Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.241533 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="config-reloader" containerID="cri-o://8106728d4589820b23f4bb38e82433fa00ceb77f299f84faa4e343985d64ece0" gracePeriod=600 Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.348733 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-r4r76-config-w9q58"] Jan 06 14:57:23 crc kubenswrapper[4744]: E0106 14:57:23.349186 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d656c5-0291-4edd-b1cd-304b1bdce02b" containerName="mariadb-database-create" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.349202 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d656c5-0291-4edd-b1cd-304b1bdce02b" containerName="mariadb-database-create" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.349405 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d656c5-0291-4edd-b1cd-304b1bdce02b" containerName="mariadb-database-create" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.350058 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.357696 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.368222 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-r4r76-config-w9q58"] Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.461545 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.461584 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr9f4\" (UniqueName: \"kubernetes.io/projected/084d0f0c-3bd0-4290-a148-3d952c51f3dd-kube-api-access-lr9f4\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.461612 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run-ovn\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.461661 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-scripts\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.461704 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-additional-scripts\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.461957 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-log-ovn\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.563786 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-log-ovn\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.563968 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.564006 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr9f4\" (UniqueName: \"kubernetes.io/projected/084d0f0c-3bd0-4290-a148-3d952c51f3dd-kube-api-access-lr9f4\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.564045 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run-ovn\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.564106 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-scripts\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.564123 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-log-ovn\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.564241 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run-ovn\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.564270 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.566073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-scripts\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.566501 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-additional-scripts\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.567247 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-additional-scripts\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.583180 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr9f4\" (UniqueName: \"kubernetes.io/projected/084d0f0c-3bd0-4290-a148-3d952c51f3dd-kube-api-access-lr9f4\") pod \"ovn-controller-r4r76-config-w9q58\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:23 crc kubenswrapper[4744]: I0106 14:57:23.676765 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:24 crc kubenswrapper[4744]: I0106 14:57:24.218305 4744 generic.go:334] "Generic (PLEG): container finished" podID="702af028-b407-4750-a0bc-40fc8f144163" containerID="a268c49ab4e371be422a1a35ca45cacd6c76688539be436d8fc478a71d2f361d" exitCode=0 Jan 06 14:57:24 crc kubenswrapper[4744]: I0106 14:57:24.218599 4744 generic.go:334] "Generic (PLEG): container finished" podID="702af028-b407-4750-a0bc-40fc8f144163" containerID="8106728d4589820b23f4bb38e82433fa00ceb77f299f84faa4e343985d64ece0" exitCode=0 Jan 06 14:57:24 crc kubenswrapper[4744]: I0106 14:57:24.218610 4744 generic.go:334] "Generic (PLEG): container finished" podID="702af028-b407-4750-a0bc-40fc8f144163" containerID="e8690c660760c008bf10dfaba0798d89793d116fd6a52d58128117656d2a976b" exitCode=0 Jan 06 14:57:24 crc kubenswrapper[4744]: I0106 14:57:24.218367 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerDied","Data":"a268c49ab4e371be422a1a35ca45cacd6c76688539be436d8fc478a71d2f361d"} Jan 06 14:57:24 crc kubenswrapper[4744]: I0106 14:57:24.218737 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerDied","Data":"8106728d4589820b23f4bb38e82433fa00ceb77f299f84faa4e343985d64ece0"} Jan 06 14:57:24 crc kubenswrapper[4744]: I0106 14:57:24.218759 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerDied","Data":"e8690c660760c008bf10dfaba0798d89793d116fd6a52d58128117656d2a976b"} Jan 06 14:57:24 crc kubenswrapper[4744]: I0106 14:57:24.428349 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-r4r76-config-w9q58"] Jan 06 14:57:25 crc kubenswrapper[4744]: W0106 14:57:25.770794 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod084d0f0c_3bd0_4290_a148_3d952c51f3dd.slice/crio-3e92ebee168d6d8a20c9de6e8bf067ef0df8795858af76bc1e712535605a2661 WatchSource:0}: Error finding container 3e92ebee168d6d8a20c9de6e8bf067ef0df8795858af76bc1e712535605a2661: Status 404 returned error can't find the container with id 3e92ebee168d6d8a20c9de6e8bf067ef0df8795858af76bc1e712535605a2661 Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.896750 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.898795 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.910601 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927470 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d7e690-64ff-4302-ab9d-57685f2abc88-operator-scripts\") pod \"b5d7e690-64ff-4302-ab9d-57685f2abc88\" (UID: \"b5d7e690-64ff-4302-ab9d-57685f2abc88\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927538 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zqt6\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-kube-api-access-8zqt6\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927657 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-thanos-prometheus-http-client-file\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927704 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-1\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927791 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/702af028-b407-4750-a0bc-40fc8f144163-config-out\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927857 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-0\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927898 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-config\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927918 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-tls-assets\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927944 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-web-config\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.927967 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxqb7\" (UniqueName: \"kubernetes.io/projected/b5d7e690-64ff-4302-ab9d-57685f2abc88-kube-api-access-gxqb7\") pod \"b5d7e690-64ff-4302-ab9d-57685f2abc88\" (UID: \"b5d7e690-64ff-4302-ab9d-57685f2abc88\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.928209 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.928264 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-2\") pod \"702af028-b407-4750-a0bc-40fc8f144163\" (UID: \"702af028-b407-4750-a0bc-40fc8f144163\") " Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.929423 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.929488 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.929994 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5d7e690-64ff-4302-ab9d-57685f2abc88-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b5d7e690-64ff-4302-ab9d-57685f2abc88" (UID: "b5d7e690-64ff-4302-ab9d-57685f2abc88"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.930459 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.943599 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.945321 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d7e690-64ff-4302-ab9d-57685f2abc88-kube-api-access-gxqb7" (OuterVolumeSpecName: "kube-api-access-gxqb7") pod "b5d7e690-64ff-4302-ab9d-57685f2abc88" (UID: "b5d7e690-64ff-4302-ab9d-57685f2abc88"). InnerVolumeSpecName "kube-api-access-gxqb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.950129 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-config" (OuterVolumeSpecName: "config") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.950654 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.954200 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-kube-api-access-8zqt6" (OuterVolumeSpecName: "kube-api-access-8zqt6") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "kube-api-access-8zqt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:25 crc kubenswrapper[4744]: I0106 14:57:25.965801 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/702af028-b407-4750-a0bc-40fc8f144163-config-out" (OuterVolumeSpecName: "config-out") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.011720 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-web-config" (OuterVolumeSpecName: "web-config") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.030975 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlw9f\" (UniqueName: \"kubernetes.io/projected/f59d0597-674a-4db8-b70c-1307186cf198-kube-api-access-rlw9f\") pod \"f59d0597-674a-4db8-b70c-1307186cf198\" (UID: \"f59d0597-674a-4db8-b70c-1307186cf198\") " Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031094 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f59d0597-674a-4db8-b70c-1307186cf198-operator-scripts\") pod \"f59d0597-674a-4db8-b70c-1307186cf198\" (UID: \"f59d0597-674a-4db8-b70c-1307186cf198\") " Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031788 4744 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/702af028-b407-4750-a0bc-40fc8f144163-config-out\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031817 4744 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031832 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031845 4744 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-tls-assets\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031856 4744 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-web-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031868 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxqb7\" (UniqueName: \"kubernetes.io/projected/b5d7e690-64ff-4302-ab9d-57685f2abc88-kube-api-access-gxqb7\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031880 4744 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031894 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b5d7e690-64ff-4302-ab9d-57685f2abc88-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031906 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zqt6\" (UniqueName: \"kubernetes.io/projected/702af028-b407-4750-a0bc-40fc8f144163-kube-api-access-8zqt6\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031921 4744 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/702af028-b407-4750-a0bc-40fc8f144163-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.031932 4744 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/702af028-b407-4750-a0bc-40fc8f144163-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.037407 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f59d0597-674a-4db8-b70c-1307186cf198-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f59d0597-674a-4db8-b70c-1307186cf198" (UID: "f59d0597-674a-4db8-b70c-1307186cf198"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.037619 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f59d0597-674a-4db8-b70c-1307186cf198-kube-api-access-rlw9f" (OuterVolumeSpecName: "kube-api-access-rlw9f") pod "f59d0597-674a-4db8-b70c-1307186cf198" (UID: "f59d0597-674a-4db8-b70c-1307186cf198"). InnerVolumeSpecName "kube-api-access-rlw9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.135050 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlw9f\" (UniqueName: \"kubernetes.io/projected/f59d0597-674a-4db8-b70c-1307186cf198-kube-api-access-rlw9f\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.135124 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f59d0597-674a-4db8-b70c-1307186cf198-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.261398 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" event={"ID":"f59d0597-674a-4db8-b70c-1307186cf198","Type":"ContainerDied","Data":"6f3aa15ea685ec06c61ed0cf793177d9fd54570da5e2385bb6a027ac3de18473"} Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.261442 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f3aa15ea685ec06c61ed0cf793177d9fd54570da5e2385bb6a027ac3de18473" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.261507 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-08dd-account-create-update-hph9l" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.268594 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "702af028-b407-4750-a0bc-40fc8f144163" (UID: "702af028-b407-4750-a0bc-40fc8f144163"). InnerVolumeSpecName "pvc-54250398-c52e-4c6a-9797-9a30efdd8220". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.274151 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8lng4" event={"ID":"b5d7e690-64ff-4302-ab9d-57685f2abc88","Type":"ContainerDied","Data":"c5a83067aacd703aa19c2b2b6efeb80b0fcd5dc4d971fba9d541bb18a5c21f06"} Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.274197 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5a83067aacd703aa19c2b2b6efeb80b0fcd5dc4d971fba9d541bb18a5c21f06" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.274248 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8lng4" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.285659 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r4r76-config-w9q58" event={"ID":"084d0f0c-3bd0-4290-a148-3d952c51f3dd","Type":"ContainerStarted","Data":"3e92ebee168d6d8a20c9de6e8bf067ef0df8795858af76bc1e712535605a2661"} Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.297280 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"702af028-b407-4750-a0bc-40fc8f144163","Type":"ContainerDied","Data":"d26a51cabb330d65bcf8efc36f345d2d133e669329ec7cd912284dd9e92b8e93"} Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.297327 4744 scope.go:117] "RemoveContainer" containerID="a268c49ab4e371be422a1a35ca45cacd6c76688539be436d8fc478a71d2f361d" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.297370 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.338487 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-54250398-c52e-4c6a-9797-9a30efdd8220\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220\") on node \"crc\" " Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.344417 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.366289 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.379272 4744 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.379461 4744 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-54250398-c52e-4c6a-9797-9a30efdd8220" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220") on node "crc" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.384269 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Jan 06 14:57:26 crc kubenswrapper[4744]: E0106 14:57:26.384693 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="thanos-sidecar" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.384712 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="thanos-sidecar" Jan 06 14:57:26 crc kubenswrapper[4744]: E0106 14:57:26.384729 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="config-reloader" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.384738 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="config-reloader" Jan 06 14:57:26 crc kubenswrapper[4744]: E0106 14:57:26.384752 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d7e690-64ff-4302-ab9d-57685f2abc88" containerName="mariadb-account-create-update" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.384758 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d7e690-64ff-4302-ab9d-57685f2abc88" containerName="mariadb-account-create-update" Jan 06 14:57:26 crc kubenswrapper[4744]: E0106 14:57:26.384779 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="init-config-reloader" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.384785 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="init-config-reloader" Jan 06 14:57:26 crc kubenswrapper[4744]: E0106 14:57:26.384793 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f59d0597-674a-4db8-b70c-1307186cf198" containerName="mariadb-account-create-update" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.384799 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f59d0597-674a-4db8-b70c-1307186cf198" containerName="mariadb-account-create-update" Jan 06 14:57:26 crc kubenswrapper[4744]: E0106 14:57:26.384824 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="prometheus" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.384831 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="prometheus" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.384996 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="config-reloader" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.385010 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="thanos-sidecar" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.385019 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d7e690-64ff-4302-ab9d-57685f2abc88" containerName="mariadb-account-create-update" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.385032 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="prometheus" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.385052 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f59d0597-674a-4db8-b70c-1307186cf198" containerName="mariadb-account-create-update" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.386726 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.391965 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.392174 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.392299 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.392496 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-f8cnq" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.392761 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.393010 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.393486 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.393619 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.403388 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.407198 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440557 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-54250398-c52e-4c6a-9797-9a30efdd8220\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440601 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km2m8\" (UniqueName: \"kubernetes.io/projected/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-kube-api-access-km2m8\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440621 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440641 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440662 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440683 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440712 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-config\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440749 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.440989 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.441024 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.441075 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.447849 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.447899 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-54250398-c52e-4c6a-9797-9a30efdd8220\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7cd4730aba1d95fd82dde750f7c410cd74c9baee4ca03dd245d4d466c88db159/globalmount\"" pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.494290 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-54250398-c52e-4c6a-9797-9a30efdd8220\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54250398-c52e-4c6a-9797-9a30efdd8220\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.542525 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km2m8\" (UniqueName: \"kubernetes.io/projected/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-kube-api-access-km2m8\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.542569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.542590 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.542608 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.542629 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.542653 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-config\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.543208 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.543238 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.543269 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.543345 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.543453 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.543936 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.543428 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.544078 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.544291 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.546583 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-config\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.546617 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.547360 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.547438 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.547481 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.548040 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.548613 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.549021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.620341 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km2m8\" (UniqueName: \"kubernetes.io/projected/0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf-kube-api-access-km2m8\") pod \"prometheus-metric-storage-0\" (UID: \"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf\") " pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:26 crc kubenswrapper[4744]: I0106 14:57:26.719711 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Jan 06 14:57:27 crc kubenswrapper[4744]: I0106 14:57:27.728237 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="702af028-b407-4750-a0bc-40fc8f144163" path="/var/lib/kubelet/pods/702af028-b407-4750-a0bc-40fc8f144163/volumes" Jan 06 14:57:27 crc kubenswrapper[4744]: I0106 14:57:27.821450 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="702af028-b407-4750-a0bc-40fc8f144163" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.137:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 14:57:27 crc kubenswrapper[4744]: I0106 14:57:27.890858 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-r4r76" podUID="7f86f187-37db-488e-b6f6-22a5becacecd" containerName="ovn-controller" probeResult="failure" output=< Jan 06 14:57:27 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 06 14:57:27 crc kubenswrapper[4744]: > Jan 06 14:57:28 crc kubenswrapper[4744]: I0106 14:57:28.401818 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 06 14:57:28 crc kubenswrapper[4744]: I0106 14:57:28.721087 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Jan 06 14:57:28 crc kubenswrapper[4744]: I0106 14:57:28.729122 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.144540 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.146123 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.150527 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.158896 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.214379 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.214433 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6flw\" (UniqueName: \"kubernetes.io/projected/fe94fea2-a713-4982-8d58-9bd7176d99ed-kube-api-access-b6flw\") pod \"mysqld-exporter-0\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.214571 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-config-data\") pod \"mysqld-exporter-0\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.316219 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-config-data\") pod \"mysqld-exporter-0\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.316307 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.316329 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6flw\" (UniqueName: \"kubernetes.io/projected/fe94fea2-a713-4982-8d58-9bd7176d99ed-kube-api-access-b6flw\") pod \"mysqld-exporter-0\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.330523 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.341819 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-config-data\") pod \"mysqld-exporter-0\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.343749 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6flw\" (UniqueName: \"kubernetes.io/projected/fe94fea2-a713-4982-8d58-9bd7176d99ed-kube-api-access-b6flw\") pod \"mysqld-exporter-0\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " pod="openstack/mysqld-exporter-0" Jan 06 14:57:29 crc kubenswrapper[4744]: I0106 14:57:29.500000 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Jan 06 14:57:33 crc kubenswrapper[4744]: I0106 14:57:33.722042 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-r4r76" podUID="7f86f187-37db-488e-b6f6-22a5becacecd" containerName="ovn-controller" probeResult="failure" output=< Jan 06 14:57:33 crc kubenswrapper[4744]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 06 14:57:33 crc kubenswrapper[4744]: > Jan 06 14:57:35 crc kubenswrapper[4744]: I0106 14:57:35.332175 4744 scope.go:117] "RemoveContainer" containerID="8106728d4589820b23f4bb38e82433fa00ceb77f299f84faa4e343985d64ece0" Jan 06 14:57:35 crc kubenswrapper[4744]: I0106 14:57:35.377099 4744 scope.go:117] "RemoveContainer" containerID="e8690c660760c008bf10dfaba0798d89793d116fd6a52d58128117656d2a976b" Jan 06 14:57:35 crc kubenswrapper[4744]: I0106 14:57:35.544390 4744 scope.go:117] "RemoveContainer" containerID="bc5b62258060bf20a1d4f1eb6935143bfd990dbab407532cfdf16ae29ab74621" Jan 06 14:57:35 crc kubenswrapper[4744]: E0106 14:57:35.639096 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Jan 06 14:57:35 crc kubenswrapper[4744]: E0106 14:57:35.639345 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hstt2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-68rr4_openstack(cb6ad61a-8190-4d4d-987c-f609c1e8cf44): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:57:35 crc kubenswrapper[4744]: E0106 14:57:35.640885 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-68rr4" podUID="cb6ad61a-8190-4d4d-987c-f609c1e8cf44" Jan 06 14:57:35 crc kubenswrapper[4744]: I0106 14:57:35.892730 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Jan 06 14:57:35 crc kubenswrapper[4744]: W0106 14:57:35.899877 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe94fea2_a713_4982_8d58_9bd7176d99ed.slice/crio-f909f832b2a68adb2c1b65b2d9c59c6cb9836d3742878a0a29ecb521d2eccbbe WatchSource:0}: Error finding container f909f832b2a68adb2c1b65b2d9c59c6cb9836d3742878a0a29ecb521d2eccbbe: Status 404 returned error can't find the container with id f909f832b2a68adb2c1b65b2d9c59c6cb9836d3742878a0a29ecb521d2eccbbe Jan 06 14:57:35 crc kubenswrapper[4744]: I0106 14:57:35.915377 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Jan 06 14:57:35 crc kubenswrapper[4744]: W0106 14:57:35.928022 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ae9e83f_2f67_419c_9929_cfd5bdcfcdbf.slice/crio-960276f9eedde9d77925f3043b1eef7af6026a81420f3c9c9cd6cfb2f7c95f8a WatchSource:0}: Error finding container 960276f9eedde9d77925f3043b1eef7af6026a81420f3c9c9cd6cfb2f7c95f8a: Status 404 returned error can't find the container with id 960276f9eedde9d77925f3043b1eef7af6026a81420f3c9c9cd6cfb2f7c95f8a Jan 06 14:57:36 crc kubenswrapper[4744]: I0106 14:57:36.407914 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r4r76-config-w9q58" event={"ID":"084d0f0c-3bd0-4290-a148-3d952c51f3dd","Type":"ContainerStarted","Data":"55c7ebd388c47561fc4ecc894a9ba12f3b56af6569453cf4d560ae54646623b3"} Jan 06 14:57:36 crc kubenswrapper[4744]: I0106 14:57:36.413345 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"b4631cb3fd08ab5e4660254f66aa073707c2795c9c31d34a40152f8d6ef0d33c"} Jan 06 14:57:36 crc kubenswrapper[4744]: I0106 14:57:36.414490 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf","Type":"ContainerStarted","Data":"960276f9eedde9d77925f3043b1eef7af6026a81420f3c9c9cd6cfb2f7c95f8a"} Jan 06 14:57:36 crc kubenswrapper[4744]: I0106 14:57:36.416090 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"fe94fea2-a713-4982-8d58-9bd7176d99ed","Type":"ContainerStarted","Data":"f909f832b2a68adb2c1b65b2d9c59c6cb9836d3742878a0a29ecb521d2eccbbe"} Jan 06 14:57:36 crc kubenswrapper[4744]: E0106 14:57:36.417239 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-68rr4" podUID="cb6ad61a-8190-4d4d-987c-f609c1e8cf44" Jan 06 14:57:36 crc kubenswrapper[4744]: I0106 14:57:36.428777 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-r4r76-config-w9q58" podStartSLOduration=13.428759971 podStartE2EDuration="13.428759971s" podCreationTimestamp="2026-01-06 14:57:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:36.42568346 +0000 UTC m=+1253.053149788" watchObservedRunningTime="2026-01-06 14:57:36.428759971 +0000 UTC m=+1253.056226289" Jan 06 14:57:37 crc kubenswrapper[4744]: I0106 14:57:37.427307 4744 generic.go:334] "Generic (PLEG): container finished" podID="084d0f0c-3bd0-4290-a148-3d952c51f3dd" containerID="55c7ebd388c47561fc4ecc894a9ba12f3b56af6569453cf4d560ae54646623b3" exitCode=0 Jan 06 14:57:37 crc kubenswrapper[4744]: I0106 14:57:37.427662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r4r76-config-w9q58" event={"ID":"084d0f0c-3bd0-4290-a148-3d952c51f3dd","Type":"ContainerDied","Data":"55c7ebd388c47561fc4ecc894a9ba12f3b56af6569453cf4d560ae54646623b3"} Jan 06 14:57:37 crc kubenswrapper[4744]: I0106 14:57:37.433087 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"b2fb1c3059944b6b60fab0bb02a956d5fbbc81a3773506559edface04d501902"} Jan 06 14:57:37 crc kubenswrapper[4744]: I0106 14:57:37.433141 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"d8b763a914397d7dbfdb29e51a4c1dcae99668d42dbd17924696527baee136d5"} Jan 06 14:57:37 crc kubenswrapper[4744]: I0106 14:57:37.892020 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-r4r76" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.722397 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.731754 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.909509 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.982354 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-log-ovn\") pod \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.982622 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "084d0f0c-3bd0-4290-a148-3d952c51f3dd" (UID: "084d0f0c-3bd0-4290-a148-3d952c51f3dd"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.982762 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr9f4\" (UniqueName: \"kubernetes.io/projected/084d0f0c-3bd0-4290-a148-3d952c51f3dd-kube-api-access-lr9f4\") pod \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.982859 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run\") pod \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.982912 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-scripts\") pod \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.982985 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run-ovn\") pod \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.983117 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-additional-scripts\") pod \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\" (UID: \"084d0f0c-3bd0-4290-a148-3d952c51f3dd\") " Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.984929 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run" (OuterVolumeSpecName: "var-run") pod "084d0f0c-3bd0-4290-a148-3d952c51f3dd" (UID: "084d0f0c-3bd0-4290-a148-3d952c51f3dd"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.985792 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "084d0f0c-3bd0-4290-a148-3d952c51f3dd" (UID: "084d0f0c-3bd0-4290-a148-3d952c51f3dd"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.986887 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "084d0f0c-3bd0-4290-a148-3d952c51f3dd" (UID: "084d0f0c-3bd0-4290-a148-3d952c51f3dd"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.987368 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-scripts" (OuterVolumeSpecName: "scripts") pod "084d0f0c-3bd0-4290-a148-3d952c51f3dd" (UID: "084d0f0c-3bd0-4290-a148-3d952c51f3dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.988454 4744 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.988527 4744 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:38 crc kubenswrapper[4744]: I0106 14:57:38.988550 4744 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/084d0f0c-3bd0-4290-a148-3d952c51f3dd-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.000463 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/084d0f0c-3bd0-4290-a148-3d952c51f3dd-kube-api-access-lr9f4" (OuterVolumeSpecName: "kube-api-access-lr9f4") pod "084d0f0c-3bd0-4290-a148-3d952c51f3dd" (UID: "084d0f0c-3bd0-4290-a148-3d952c51f3dd"). InnerVolumeSpecName "kube-api-access-lr9f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.091551 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.091596 4744 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/084d0f0c-3bd0-4290-a148-3d952c51f3dd-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.091612 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr9f4\" (UniqueName: \"kubernetes.io/projected/084d0f0c-3bd0-4290-a148-3d952c51f3dd-kube-api-access-lr9f4\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.458680 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-r4r76-config-w9q58" event={"ID":"084d0f0c-3bd0-4290-a148-3d952c51f3dd","Type":"ContainerDied","Data":"3e92ebee168d6d8a20c9de6e8bf067ef0df8795858af76bc1e712535605a2661"} Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.458728 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e92ebee168d6d8a20c9de6e8bf067ef0df8795858af76bc1e712535605a2661" Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.458705 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-r4r76-config-w9q58" Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.463820 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"d82c3a28857485b6cfe30daabd9acf441f3c6fa754d7c71286ff140db8901e3d"} Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.552370 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-r4r76-config-w9q58"] Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.569777 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-r4r76-config-w9q58"] Jan 06 14:57:39 crc kubenswrapper[4744]: I0106 14:57:39.735364 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="084d0f0c-3bd0-4290-a148-3d952c51f3dd" path="/var/lib/kubelet/pods/084d0f0c-3bd0-4290-a148-3d952c51f3dd/volumes" Jan 06 14:57:40 crc kubenswrapper[4744]: I0106 14:57:40.480061 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf","Type":"ContainerStarted","Data":"e98b63d3ec96f756dcbed447564b013bcd722583963ce64dddb6a2fe9221dbfa"} Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.490030 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-79c64"] Jan 06 14:57:41 crc kubenswrapper[4744]: E0106 14:57:41.491960 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="084d0f0c-3bd0-4290-a148-3d952c51f3dd" containerName="ovn-config" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.492036 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="084d0f0c-3bd0-4290-a148-3d952c51f3dd" containerName="ovn-config" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.492394 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="084d0f0c-3bd0-4290-a148-3d952c51f3dd" containerName="ovn-config" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.493896 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-79c64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.507715 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-79c64"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.546755 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16de9a20-7669-4935-9995-a5d1172e68e3-operator-scripts\") pod \"cinder-db-create-79c64\" (UID: \"16de9a20-7669-4935-9995-a5d1172e68e3\") " pod="openstack/cinder-db-create-79c64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.546899 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8jmk\" (UniqueName: \"kubernetes.io/projected/16de9a20-7669-4935-9995-a5d1172e68e3-kube-api-access-l8jmk\") pod \"cinder-db-create-79c64\" (UID: \"16de9a20-7669-4935-9995-a5d1172e68e3\") " pod="openstack/cinder-db-create-79c64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.586309 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-s8j7d"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.588179 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.598884 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-s8j7d"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.649780 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16de9a20-7669-4935-9995-a5d1172e68e3-operator-scripts\") pod \"cinder-db-create-79c64\" (UID: \"16de9a20-7669-4935-9995-a5d1172e68e3\") " pod="openstack/cinder-db-create-79c64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.650196 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8jmk\" (UniqueName: \"kubernetes.io/projected/16de9a20-7669-4935-9995-a5d1172e68e3-kube-api-access-l8jmk\") pod \"cinder-db-create-79c64\" (UID: \"16de9a20-7669-4935-9995-a5d1172e68e3\") " pod="openstack/cinder-db-create-79c64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.650279 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afdb36a9-143a-4bbc-a826-cdfa078e5d68-operator-scripts\") pod \"barbican-db-create-s8j7d\" (UID: \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\") " pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.650458 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt6sq\" (UniqueName: \"kubernetes.io/projected/afdb36a9-143a-4bbc-a826-cdfa078e5d68-kube-api-access-kt6sq\") pod \"barbican-db-create-s8j7d\" (UID: \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\") " pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.650686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16de9a20-7669-4935-9995-a5d1172e68e3-operator-scripts\") pod \"cinder-db-create-79c64\" (UID: \"16de9a20-7669-4935-9995-a5d1172e68e3\") " pod="openstack/cinder-db-create-79c64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.676080 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8jmk\" (UniqueName: \"kubernetes.io/projected/16de9a20-7669-4935-9995-a5d1172e68e3-kube-api-access-l8jmk\") pod \"cinder-db-create-79c64\" (UID: \"16de9a20-7669-4935-9995-a5d1172e68e3\") " pod="openstack/cinder-db-create-79c64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.692785 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f033-account-create-update-w89dh"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.694454 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.698465 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.704516 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f033-account-create-update-w89dh"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.754600 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afdb36a9-143a-4bbc-a826-cdfa078e5d68-operator-scripts\") pod \"barbican-db-create-s8j7d\" (UID: \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\") " pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.754694 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt6sq\" (UniqueName: \"kubernetes.io/projected/afdb36a9-143a-4bbc-a826-cdfa078e5d68-kube-api-access-kt6sq\") pod \"barbican-db-create-s8j7d\" (UID: \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\") " pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.754738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7l7r\" (UniqueName: \"kubernetes.io/projected/a372a06b-1453-4141-b82b-96af30ceb608-kube-api-access-k7l7r\") pod \"cinder-f033-account-create-update-w89dh\" (UID: \"a372a06b-1453-4141-b82b-96af30ceb608\") " pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.754789 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a372a06b-1453-4141-b82b-96af30ceb608-operator-scripts\") pod \"cinder-f033-account-create-update-w89dh\" (UID: \"a372a06b-1453-4141-b82b-96af30ceb608\") " pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.755535 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afdb36a9-143a-4bbc-a826-cdfa078e5d68-operator-scripts\") pod \"barbican-db-create-s8j7d\" (UID: \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\") " pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.780387 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-gwmvk"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.782786 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.784769 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt6sq\" (UniqueName: \"kubernetes.io/projected/afdb36a9-143a-4bbc-a826-cdfa078e5d68-kube-api-access-kt6sq\") pod \"barbican-db-create-s8j7d\" (UID: \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\") " pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.810629 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-f5a5-account-create-update-jkh64"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.817710 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-79c64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.832965 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-gwmvk"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.833339 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.834062 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-f5a5-account-create-update-jkh64"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.838613 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.859142 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7l7r\" (UniqueName: \"kubernetes.io/projected/a372a06b-1453-4141-b82b-96af30ceb608-kube-api-access-k7l7r\") pod \"cinder-f033-account-create-update-w89dh\" (UID: \"a372a06b-1453-4141-b82b-96af30ceb608\") " pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.862786 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a372a06b-1453-4141-b82b-96af30ceb608-operator-scripts\") pod \"cinder-f033-account-create-update-w89dh\" (UID: \"a372a06b-1453-4141-b82b-96af30ceb608\") " pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.873243 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a372a06b-1453-4141-b82b-96af30ceb608-operator-scripts\") pod \"cinder-f033-account-create-update-w89dh\" (UID: \"a372a06b-1453-4141-b82b-96af30ceb608\") " pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.919974 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7l7r\" (UniqueName: \"kubernetes.io/projected/a372a06b-1453-4141-b82b-96af30ceb608-kube-api-access-k7l7r\") pod \"cinder-f033-account-create-update-w89dh\" (UID: \"a372a06b-1453-4141-b82b-96af30ceb608\") " pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.933256 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.940952 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-jbwmt"] Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.946712 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.952384 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.953339 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.953381 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.953461 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8hqr4" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.968447 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6cbs\" (UniqueName: \"kubernetes.io/projected/44b3ba14-81e9-4338-bc3e-297128ea5ef1-kube-api-access-m6cbs\") pod \"heat-f5a5-account-create-update-jkh64\" (UID: \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\") " pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.968523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc6mt\" (UniqueName: \"kubernetes.io/projected/a57dc900-6506-4a69-b49a-3bd5dd6a88db-kube-api-access-lc6mt\") pod \"heat-db-create-gwmvk\" (UID: \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\") " pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.968675 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57dc900-6506-4a69-b49a-3bd5dd6a88db-operator-scripts\") pod \"heat-db-create-gwmvk\" (UID: \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\") " pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:41 crc kubenswrapper[4744]: I0106 14:57:41.968703 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44b3ba14-81e9-4338-bc3e-297128ea5ef1-operator-scripts\") pod \"heat-f5a5-account-create-update-jkh64\" (UID: \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\") " pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.004561 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jbwmt"] Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.016592 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-6252-account-create-update-cz7bz"] Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.017966 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.022385 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.029656 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6252-account-create-update-cz7bz"] Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.038566 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.092483 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-config-data\") pod \"keystone-db-sync-jbwmt\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.092531 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57dc900-6506-4a69-b49a-3bd5dd6a88db-operator-scripts\") pod \"heat-db-create-gwmvk\" (UID: \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\") " pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.092563 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44b3ba14-81e9-4338-bc3e-297128ea5ef1-operator-scripts\") pod \"heat-f5a5-account-create-update-jkh64\" (UID: \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\") " pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.092588 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41cf511d-549f-4b68-bb95-40a73cddbf83-operator-scripts\") pod \"barbican-6252-account-create-update-cz7bz\" (UID: \"41cf511d-549f-4b68-bb95-40a73cddbf83\") " pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.092636 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-combined-ca-bundle\") pod \"keystone-db-sync-jbwmt\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.092666 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6cbs\" (UniqueName: \"kubernetes.io/projected/44b3ba14-81e9-4338-bc3e-297128ea5ef1-kube-api-access-m6cbs\") pod \"heat-f5a5-account-create-update-jkh64\" (UID: \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\") " pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.092704 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc6mt\" (UniqueName: \"kubernetes.io/projected/a57dc900-6506-4a69-b49a-3bd5dd6a88db-kube-api-access-lc6mt\") pod \"heat-db-create-gwmvk\" (UID: \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\") " pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.092743 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbnbf\" (UniqueName: \"kubernetes.io/projected/41cf511d-549f-4b68-bb95-40a73cddbf83-kube-api-access-cbnbf\") pod \"barbican-6252-account-create-update-cz7bz\" (UID: \"41cf511d-549f-4b68-bb95-40a73cddbf83\") " pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.092782 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx8dp\" (UniqueName: \"kubernetes.io/projected/66f0fe7a-9061-4bc2-a957-f9d2926401d7-kube-api-access-mx8dp\") pod \"keystone-db-sync-jbwmt\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.093409 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57dc900-6506-4a69-b49a-3bd5dd6a88db-operator-scripts\") pod \"heat-db-create-gwmvk\" (UID: \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\") " pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.093614 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44b3ba14-81e9-4338-bc3e-297128ea5ef1-operator-scripts\") pod \"heat-f5a5-account-create-update-jkh64\" (UID: \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\") " pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.134543 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc6mt\" (UniqueName: \"kubernetes.io/projected/a57dc900-6506-4a69-b49a-3bd5dd6a88db-kube-api-access-lc6mt\") pod \"heat-db-create-gwmvk\" (UID: \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\") " pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.135508 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.150752 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6cbs\" (UniqueName: \"kubernetes.io/projected/44b3ba14-81e9-4338-bc3e-297128ea5ef1-kube-api-access-m6cbs\") pod \"heat-f5a5-account-create-update-jkh64\" (UID: \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\") " pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.189442 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.196243 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-combined-ca-bundle\") pod \"keystone-db-sync-jbwmt\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.196345 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbnbf\" (UniqueName: \"kubernetes.io/projected/41cf511d-549f-4b68-bb95-40a73cddbf83-kube-api-access-cbnbf\") pod \"barbican-6252-account-create-update-cz7bz\" (UID: \"41cf511d-549f-4b68-bb95-40a73cddbf83\") " pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.196383 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx8dp\" (UniqueName: \"kubernetes.io/projected/66f0fe7a-9061-4bc2-a957-f9d2926401d7-kube-api-access-mx8dp\") pod \"keystone-db-sync-jbwmt\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.196440 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-config-data\") pod \"keystone-db-sync-jbwmt\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.196489 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41cf511d-549f-4b68-bb95-40a73cddbf83-operator-scripts\") pod \"barbican-6252-account-create-update-cz7bz\" (UID: \"41cf511d-549f-4b68-bb95-40a73cddbf83\") " pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.197285 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41cf511d-549f-4b68-bb95-40a73cddbf83-operator-scripts\") pod \"barbican-6252-account-create-update-cz7bz\" (UID: \"41cf511d-549f-4b68-bb95-40a73cddbf83\") " pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.209767 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-combined-ca-bundle\") pod \"keystone-db-sync-jbwmt\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.212399 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-config-data\") pod \"keystone-db-sync-jbwmt\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.246531 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbnbf\" (UniqueName: \"kubernetes.io/projected/41cf511d-549f-4b68-bb95-40a73cddbf83-kube-api-access-cbnbf\") pod \"barbican-6252-account-create-update-cz7bz\" (UID: \"41cf511d-549f-4b68-bb95-40a73cddbf83\") " pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.246544 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx8dp\" (UniqueName: \"kubernetes.io/projected/66f0fe7a-9061-4bc2-a957-f9d2926401d7-kube-api-access-mx8dp\") pod \"keystone-db-sync-jbwmt\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.253289 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.286774 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.614999 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-79c64"] Jan 06 14:57:42 crc kubenswrapper[4744]: I0106 14:57:42.926526 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-s8j7d"] Jan 06 14:57:42 crc kubenswrapper[4744]: W0106 14:57:42.945343 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafdb36a9_143a_4bbc_a826_cdfa078e5d68.slice/crio-5fa163d2ec0b65249a921f9bde5486ca6fe1040815ea1a682da60036045cc628 WatchSource:0}: Error finding container 5fa163d2ec0b65249a921f9bde5486ca6fe1040815ea1a682da60036045cc628: Status 404 returned error can't find the container with id 5fa163d2ec0b65249a921f9bde5486ca6fe1040815ea1a682da60036045cc628 Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.051578 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f033-account-create-update-w89dh"] Jan 06 14:57:43 crc kubenswrapper[4744]: W0106 14:57:43.054565 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda372a06b_1453_4141_b82b_96af30ceb608.slice/crio-b59300fe344318d36f9e3c383cbc923fb3e949f53be71097096dddf6ca49c280 WatchSource:0}: Error finding container b59300fe344318d36f9e3c383cbc923fb3e949f53be71097096dddf6ca49c280: Status 404 returned error can't find the container with id b59300fe344318d36f9e3c383cbc923fb3e949f53be71097096dddf6ca49c280 Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.179332 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-gwmvk"] Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.191511 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-f5a5-account-create-update-jkh64"] Jan 06 14:57:43 crc kubenswrapper[4744]: W0106 14:57:43.193731 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda57dc900_6506_4a69_b49a_3bd5dd6a88db.slice/crio-88b8ac24a6860c9c2ada49f919f87ed6921ec7767e67258f496de3a82489f0b9 WatchSource:0}: Error finding container 88b8ac24a6860c9c2ada49f919f87ed6921ec7767e67258f496de3a82489f0b9: Status 404 returned error can't find the container with id 88b8ac24a6860c9c2ada49f919f87ed6921ec7767e67258f496de3a82489f0b9 Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.345234 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-jbwmt"] Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.466810 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-6252-account-create-update-cz7bz"] Jan 06 14:57:43 crc kubenswrapper[4744]: W0106 14:57:43.468997 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41cf511d_549f_4b68_bb95_40a73cddbf83.slice/crio-41d4391bb14a7af6adff42223e866affd19a1ccc22691ec72d00e557a8fc6a31 WatchSource:0}: Error finding container 41d4391bb14a7af6adff42223e866affd19a1ccc22691ec72d00e557a8fc6a31: Status 404 returned error can't find the container with id 41d4391bb14a7af6adff42223e866affd19a1ccc22691ec72d00e557a8fc6a31 Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.552732 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f033-account-create-update-w89dh" event={"ID":"a372a06b-1453-4141-b82b-96af30ceb608","Type":"ContainerStarted","Data":"b59300fe344318d36f9e3c383cbc923fb3e949f53be71097096dddf6ca49c280"} Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.553973 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gwmvk" event={"ID":"a57dc900-6506-4a69-b49a-3bd5dd6a88db","Type":"ContainerStarted","Data":"88b8ac24a6860c9c2ada49f919f87ed6921ec7767e67258f496de3a82489f0b9"} Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.555760 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jbwmt" event={"ID":"66f0fe7a-9061-4bc2-a957-f9d2926401d7","Type":"ContainerStarted","Data":"58e753163e29654a79be41d8f1635fbc872d800b8b3743fff7e6db8e7cf4f18d"} Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.558081 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6252-account-create-update-cz7bz" event={"ID":"41cf511d-549f-4b68-bb95-40a73cddbf83","Type":"ContainerStarted","Data":"41d4391bb14a7af6adff42223e866affd19a1ccc22691ec72d00e557a8fc6a31"} Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.559715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-79c64" event={"ID":"16de9a20-7669-4935-9995-a5d1172e68e3","Type":"ContainerStarted","Data":"c91a28f230d3a7499551580354ab0c02c3a62ed561417134c29f27e69baf8706"} Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.560895 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s8j7d" event={"ID":"afdb36a9-143a-4bbc-a826-cdfa078e5d68","Type":"ContainerStarted","Data":"5fa163d2ec0b65249a921f9bde5486ca6fe1040815ea1a682da60036045cc628"} Jan 06 14:57:43 crc kubenswrapper[4744]: I0106 14:57:43.562212 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-f5a5-account-create-update-jkh64" event={"ID":"44b3ba14-81e9-4338-bc3e-297128ea5ef1","Type":"ContainerStarted","Data":"5916455359c55fd9df8ea03f6559e26a22b5177ca2c3985a192313ddb51f3cde"} Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.424074 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.424386 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.500451 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-67ffc"] Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.503123 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.535148 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-67ffc"] Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.606932 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"fe94fea2-a713-4982-8d58-9bd7176d99ed","Type":"ContainerStarted","Data":"19395dd7071ace70944a6c11b1d8ecdf40f737a05160e701afe0a94fdc68ea90"} Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.681532 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/795cf078-881c-4373-9e3c-13afd8662252-operator-scripts\") pod \"neutron-db-create-67ffc\" (UID: \"795cf078-881c-4373-9e3c-13afd8662252\") " pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.681648 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhjqt\" (UniqueName: \"kubernetes.io/projected/795cf078-881c-4373-9e3c-13afd8662252-kube-api-access-bhjqt\") pod \"neutron-db-create-67ffc\" (UID: \"795cf078-881c-4373-9e3c-13afd8662252\") " pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.700499 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6b41-account-create-update-9qd64"] Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.702154 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.705502 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.716934 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6b41-account-create-update-9qd64"] Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.783749 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhjqt\" (UniqueName: \"kubernetes.io/projected/795cf078-881c-4373-9e3c-13afd8662252-kube-api-access-bhjqt\") pod \"neutron-db-create-67ffc\" (UID: \"795cf078-881c-4373-9e3c-13afd8662252\") " pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.784059 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/795cf078-881c-4373-9e3c-13afd8662252-operator-scripts\") pod \"neutron-db-create-67ffc\" (UID: \"795cf078-881c-4373-9e3c-13afd8662252\") " pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.784959 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/795cf078-881c-4373-9e3c-13afd8662252-operator-scripts\") pod \"neutron-db-create-67ffc\" (UID: \"795cf078-881c-4373-9e3c-13afd8662252\") " pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.804284 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhjqt\" (UniqueName: \"kubernetes.io/projected/795cf078-881c-4373-9e3c-13afd8662252-kube-api-access-bhjqt\") pod \"neutron-db-create-67ffc\" (UID: \"795cf078-881c-4373-9e3c-13afd8662252\") " pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.836109 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.886073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxvcm\" (UniqueName: \"kubernetes.io/projected/5df8f82c-aba7-4d44-b4bd-c3a616afc480-kube-api-access-rxvcm\") pod \"neutron-6b41-account-create-update-9qd64\" (UID: \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\") " pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.886400 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5df8f82c-aba7-4d44-b4bd-c3a616afc480-operator-scripts\") pod \"neutron-6b41-account-create-update-9qd64\" (UID: \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\") " pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.990892 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxvcm\" (UniqueName: \"kubernetes.io/projected/5df8f82c-aba7-4d44-b4bd-c3a616afc480-kube-api-access-rxvcm\") pod \"neutron-6b41-account-create-update-9qd64\" (UID: \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\") " pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.991053 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5df8f82c-aba7-4d44-b4bd-c3a616afc480-operator-scripts\") pod \"neutron-6b41-account-create-update-9qd64\" (UID: \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\") " pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:44 crc kubenswrapper[4744]: I0106 14:57:44.992336 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5df8f82c-aba7-4d44-b4bd-c3a616afc480-operator-scripts\") pod \"neutron-6b41-account-create-update-9qd64\" (UID: \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\") " pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.020855 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxvcm\" (UniqueName: \"kubernetes.io/projected/5df8f82c-aba7-4d44-b4bd-c3a616afc480-kube-api-access-rxvcm\") pod \"neutron-6b41-account-create-update-9qd64\" (UID: \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\") " pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.021392 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.644623 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gwmvk" event={"ID":"a57dc900-6506-4a69-b49a-3bd5dd6a88db","Type":"ContainerStarted","Data":"10afb9bf4e76dc90e94ac82aeb60ff829b277ed81be73aaa635dc23ca0e5be73"} Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.660404 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-79c64" event={"ID":"16de9a20-7669-4935-9995-a5d1172e68e3","Type":"ContainerStarted","Data":"788bcdfaf4046f9a2e160764587f4eb09b4bcbb1d0ea440afaf7d4d09290f34a"} Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.668287 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-f5a5-account-create-update-jkh64" event={"ID":"44b3ba14-81e9-4338-bc3e-297128ea5ef1","Type":"ContainerStarted","Data":"ee305d5dfd8895fcbdc3e6f78124e6aab08b28f9b844ec202b6b0f4730d2722b"} Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.673713 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f033-account-create-update-w89dh" event={"ID":"a372a06b-1453-4141-b82b-96af30ceb608","Type":"ContainerStarted","Data":"89c738924cdbe81aff4c744ecbdb2598b0a9cb084e30bfe98239d5830b404c4a"} Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.680903 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s8j7d" event={"ID":"afdb36a9-143a-4bbc-a826-cdfa078e5d68","Type":"ContainerStarted","Data":"d27c59d832129edb676e4a2f4b297dcfbc8077bc9e58589178053d5fec161fc2"} Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.691620 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6252-account-create-update-cz7bz" event={"ID":"41cf511d-549f-4b68-bb95-40a73cddbf83","Type":"ContainerStarted","Data":"86c99a2279059f4f70da24573ab4662daad672dc4a447bc5d4cb0b1a275f468b"} Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.697246 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-67ffc"] Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.701505 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-gwmvk" podStartSLOduration=4.701471784 podStartE2EDuration="4.701471784s" podCreationTimestamp="2026-01-06 14:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:45.67562379 +0000 UTC m=+1262.303090108" watchObservedRunningTime="2026-01-06 14:57:45.701471784 +0000 UTC m=+1262.328938102" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.730519 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-79c64" podStartSLOduration=4.730470481 podStartE2EDuration="4.730470481s" podCreationTimestamp="2026-01-06 14:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:45.691087469 +0000 UTC m=+1262.318553797" watchObservedRunningTime="2026-01-06 14:57:45.730470481 +0000 UTC m=+1262.357936809" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.746625 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-f033-account-create-update-w89dh" podStartSLOduration=4.746598077 podStartE2EDuration="4.746598077s" podCreationTimestamp="2026-01-06 14:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:45.709298381 +0000 UTC m=+1262.336764699" watchObservedRunningTime="2026-01-06 14:57:45.746598077 +0000 UTC m=+1262.374064395" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.829471 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-f5a5-account-create-update-jkh64" podStartSLOduration=4.829444928 podStartE2EDuration="4.829444928s" podCreationTimestamp="2026-01-06 14:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:45.726528956 +0000 UTC m=+1262.353995274" watchObservedRunningTime="2026-01-06 14:57:45.829444928 +0000 UTC m=+1262.456911246" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.855718 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-6252-account-create-update-cz7bz" podStartSLOduration=4.855680862 podStartE2EDuration="4.855680862s" podCreationTimestamp="2026-01-06 14:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:45.752214886 +0000 UTC m=+1262.379681204" watchObservedRunningTime="2026-01-06 14:57:45.855680862 +0000 UTC m=+1262.483147180" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.884515 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=12.039759367 podStartE2EDuration="16.884487584s" podCreationTimestamp="2026-01-06 14:57:29 +0000 UTC" firstStartedPulling="2026-01-06 14:57:35.902291859 +0000 UTC m=+1252.529758177" lastFinishedPulling="2026-01-06 14:57:40.747020076 +0000 UTC m=+1257.374486394" observedRunningTime="2026-01-06 14:57:45.771550957 +0000 UTC m=+1262.399017275" watchObservedRunningTime="2026-01-06 14:57:45.884487584 +0000 UTC m=+1262.511953902" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.904434 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-s8j7d" podStartSLOduration=4.90438607 podStartE2EDuration="4.90438607s" podCreationTimestamp="2026-01-06 14:57:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:45.798084509 +0000 UTC m=+1262.425550857" watchObservedRunningTime="2026-01-06 14:57:45.90438607 +0000 UTC m=+1262.531852388" Jan 06 14:57:45 crc kubenswrapper[4744]: I0106 14:57:45.943696 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6b41-account-create-update-9qd64"] Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.711270 4744 generic.go:334] "Generic (PLEG): container finished" podID="a57dc900-6506-4a69-b49a-3bd5dd6a88db" containerID="10afb9bf4e76dc90e94ac82aeb60ff829b277ed81be73aaa635dc23ca0e5be73" exitCode=0 Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.711336 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gwmvk" event={"ID":"a57dc900-6506-4a69-b49a-3bd5dd6a88db","Type":"ContainerDied","Data":"10afb9bf4e76dc90e94ac82aeb60ff829b277ed81be73aaa635dc23ca0e5be73"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.718438 4744 generic.go:334] "Generic (PLEG): container finished" podID="41cf511d-549f-4b68-bb95-40a73cddbf83" containerID="86c99a2279059f4f70da24573ab4662daad672dc4a447bc5d4cb0b1a275f468b" exitCode=0 Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.718506 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6252-account-create-update-cz7bz" event={"ID":"41cf511d-549f-4b68-bb95-40a73cddbf83","Type":"ContainerDied","Data":"86c99a2279059f4f70da24573ab4662daad672dc4a447bc5d4cb0b1a275f468b"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.720849 4744 generic.go:334] "Generic (PLEG): container finished" podID="16de9a20-7669-4935-9995-a5d1172e68e3" containerID="788bcdfaf4046f9a2e160764587f4eb09b4bcbb1d0ea440afaf7d4d09290f34a" exitCode=0 Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.721001 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-79c64" event={"ID":"16de9a20-7669-4935-9995-a5d1172e68e3","Type":"ContainerDied","Data":"788bcdfaf4046f9a2e160764587f4eb09b4bcbb1d0ea440afaf7d4d09290f34a"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.723577 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b41-account-create-update-9qd64" event={"ID":"5df8f82c-aba7-4d44-b4bd-c3a616afc480","Type":"ContainerStarted","Data":"20e35d14624d384244c963aa5a35b1c9117ccb44e1174e5ce9d5eb1e281f03f1"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.723608 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b41-account-create-update-9qd64" event={"ID":"5df8f82c-aba7-4d44-b4bd-c3a616afc480","Type":"ContainerStarted","Data":"dd540dd776910fbf7153ed633554ce2d09bb5ecf08cd93818294595242962a6e"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.729568 4744 generic.go:334] "Generic (PLEG): container finished" podID="afdb36a9-143a-4bbc-a826-cdfa078e5d68" containerID="d27c59d832129edb676e4a2f4b297dcfbc8077bc9e58589178053d5fec161fc2" exitCode=0 Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.729657 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s8j7d" event={"ID":"afdb36a9-143a-4bbc-a826-cdfa078e5d68","Type":"ContainerDied","Data":"d27c59d832129edb676e4a2f4b297dcfbc8077bc9e58589178053d5fec161fc2"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.734257 4744 generic.go:334] "Generic (PLEG): container finished" podID="44b3ba14-81e9-4338-bc3e-297128ea5ef1" containerID="ee305d5dfd8895fcbdc3e6f78124e6aab08b28f9b844ec202b6b0f4730d2722b" exitCode=0 Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.734309 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-f5a5-account-create-update-jkh64" event={"ID":"44b3ba14-81e9-4338-bc3e-297128ea5ef1","Type":"ContainerDied","Data":"ee305d5dfd8895fcbdc3e6f78124e6aab08b28f9b844ec202b6b0f4730d2722b"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.737301 4744 generic.go:334] "Generic (PLEG): container finished" podID="795cf078-881c-4373-9e3c-13afd8662252" containerID="ce294c65bfe21d8d232f8241b1fb908d05a276361b0c60a5be9601443075bf98" exitCode=0 Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.737390 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-67ffc" event={"ID":"795cf078-881c-4373-9e3c-13afd8662252","Type":"ContainerDied","Data":"ce294c65bfe21d8d232f8241b1fb908d05a276361b0c60a5be9601443075bf98"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.737422 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-67ffc" event={"ID":"795cf078-881c-4373-9e3c-13afd8662252","Type":"ContainerStarted","Data":"bb549809ef0f9ae7928825bf80dc4e1375e6bacb8a01173d62076b0e6afad022"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.740851 4744 generic.go:334] "Generic (PLEG): container finished" podID="a372a06b-1453-4141-b82b-96af30ceb608" containerID="89c738924cdbe81aff4c744ecbdb2598b0a9cb084e30bfe98239d5830b404c4a" exitCode=0 Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.740906 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f033-account-create-update-w89dh" event={"ID":"a372a06b-1453-4141-b82b-96af30ceb608","Type":"ContainerDied","Data":"89c738924cdbe81aff4c744ecbdb2598b0a9cb084e30bfe98239d5830b404c4a"} Jan 06 14:57:46 crc kubenswrapper[4744]: I0106 14:57:46.798456 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6b41-account-create-update-9qd64" podStartSLOduration=2.798438103 podStartE2EDuration="2.798438103s" podCreationTimestamp="2026-01-06 14:57:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:57:46.791661694 +0000 UTC m=+1263.419128032" watchObservedRunningTime="2026-01-06 14:57:46.798438103 +0000 UTC m=+1263.425904421" Jan 06 14:57:47 crc kubenswrapper[4744]: I0106 14:57:47.765265 4744 generic.go:334] "Generic (PLEG): container finished" podID="5df8f82c-aba7-4d44-b4bd-c3a616afc480" containerID="20e35d14624d384244c963aa5a35b1c9117ccb44e1174e5ce9d5eb1e281f03f1" exitCode=0 Jan 06 14:57:47 crc kubenswrapper[4744]: I0106 14:57:47.765485 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b41-account-create-update-9qd64" event={"ID":"5df8f82c-aba7-4d44-b4bd-c3a616afc480","Type":"ContainerDied","Data":"20e35d14624d384244c963aa5a35b1c9117ccb44e1174e5ce9d5eb1e281f03f1"} Jan 06 14:57:47 crc kubenswrapper[4744]: I0106 14:57:47.771712 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"e50aa33fca25d70c341961f049f106c3d7b9a163bacaaf008fc223b35c8ee2ea"} Jan 06 14:57:47 crc kubenswrapper[4744]: I0106 14:57:47.771769 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"90bd787182fe69f132591a9b11b4b73b1fbaa70c930a9e73c821d136bf8fb0d4"} Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.885329 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-67ffc" event={"ID":"795cf078-881c-4373-9e3c-13afd8662252","Type":"ContainerDied","Data":"bb549809ef0f9ae7928825bf80dc4e1375e6bacb8a01173d62076b0e6afad022"} Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.886427 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb549809ef0f9ae7928825bf80dc4e1375e6bacb8a01173d62076b0e6afad022" Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.891724 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-f5a5-account-create-update-jkh64" event={"ID":"44b3ba14-81e9-4338-bc3e-297128ea5ef1","Type":"ContainerDied","Data":"5916455359c55fd9df8ea03f6559e26a22b5177ca2c3985a192313ddb51f3cde"} Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.891814 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5916455359c55fd9df8ea03f6559e26a22b5177ca2c3985a192313ddb51f3cde" Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.894716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f033-account-create-update-w89dh" event={"ID":"a372a06b-1453-4141-b82b-96af30ceb608","Type":"ContainerDied","Data":"b59300fe344318d36f9e3c383cbc923fb3e949f53be71097096dddf6ca49c280"} Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.894774 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b59300fe344318d36f9e3c383cbc923fb3e949f53be71097096dddf6ca49c280" Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.898081 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-gwmvk" event={"ID":"a57dc900-6506-4a69-b49a-3bd5dd6a88db","Type":"ContainerDied","Data":"88b8ac24a6860c9c2ada49f919f87ed6921ec7767e67258f496de3a82489f0b9"} Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.898118 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88b8ac24a6860c9c2ada49f919f87ed6921ec7767e67258f496de3a82489f0b9" Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.903094 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-6252-account-create-update-cz7bz" event={"ID":"41cf511d-549f-4b68-bb95-40a73cddbf83","Type":"ContainerDied","Data":"41d4391bb14a7af6adff42223e866affd19a1ccc22691ec72d00e557a8fc6a31"} Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.903150 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41d4391bb14a7af6adff42223e866affd19a1ccc22691ec72d00e557a8fc6a31" Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.905390 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-79c64" event={"ID":"16de9a20-7669-4935-9995-a5d1172e68e3","Type":"ContainerDied","Data":"c91a28f230d3a7499551580354ab0c02c3a62ed561417134c29f27e69baf8706"} Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.905445 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c91a28f230d3a7499551580354ab0c02c3a62ed561417134c29f27e69baf8706" Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.908224 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6b41-account-create-update-9qd64" event={"ID":"5df8f82c-aba7-4d44-b4bd-c3a616afc480","Type":"ContainerDied","Data":"dd540dd776910fbf7153ed633554ce2d09bb5ecf08cd93818294595242962a6e"} Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.908280 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd540dd776910fbf7153ed633554ce2d09bb5ecf08cd93818294595242962a6e" Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.910054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-s8j7d" event={"ID":"afdb36a9-143a-4bbc-a826-cdfa078e5d68","Type":"ContainerDied","Data":"5fa163d2ec0b65249a921f9bde5486ca6fe1040815ea1a682da60036045cc628"} Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.910104 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fa163d2ec0b65249a921f9bde5486ca6fe1040815ea1a682da60036045cc628" Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.980890 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:54 crc kubenswrapper[4744]: I0106 14:57:54.990044 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.073679 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.084427 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.107346 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44b3ba14-81e9-4338-bc3e-297128ea5ef1-operator-scripts\") pod \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\" (UID: \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.107404 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/795cf078-881c-4373-9e3c-13afd8662252-operator-scripts\") pod \"795cf078-881c-4373-9e3c-13afd8662252\" (UID: \"795cf078-881c-4373-9e3c-13afd8662252\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.107486 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6cbs\" (UniqueName: \"kubernetes.io/projected/44b3ba14-81e9-4338-bc3e-297128ea5ef1-kube-api-access-m6cbs\") pod \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\" (UID: \"44b3ba14-81e9-4338-bc3e-297128ea5ef1\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.107522 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhjqt\" (UniqueName: \"kubernetes.io/projected/795cf078-881c-4373-9e3c-13afd8662252-kube-api-access-bhjqt\") pod \"795cf078-881c-4373-9e3c-13afd8662252\" (UID: \"795cf078-881c-4373-9e3c-13afd8662252\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.108428 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/795cf078-881c-4373-9e3c-13afd8662252-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "795cf078-881c-4373-9e3c-13afd8662252" (UID: "795cf078-881c-4373-9e3c-13afd8662252"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.108471 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44b3ba14-81e9-4338-bc3e-297128ea5ef1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "44b3ba14-81e9-4338-bc3e-297128ea5ef1" (UID: "44b3ba14-81e9-4338-bc3e-297128ea5ef1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.109074 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/795cf078-881c-4373-9e3c-13afd8662252-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.109102 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44b3ba14-81e9-4338-bc3e-297128ea5ef1-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.111696 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/795cf078-881c-4373-9e3c-13afd8662252-kube-api-access-bhjqt" (OuterVolumeSpecName: "kube-api-access-bhjqt") pod "795cf078-881c-4373-9e3c-13afd8662252" (UID: "795cf078-881c-4373-9e3c-13afd8662252"). InnerVolumeSpecName "kube-api-access-bhjqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.112991 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b3ba14-81e9-4338-bc3e-297128ea5ef1-kube-api-access-m6cbs" (OuterVolumeSpecName: "kube-api-access-m6cbs") pod "44b3ba14-81e9-4338-bc3e-297128ea5ef1" (UID: "44b3ba14-81e9-4338-bc3e-297128ea5ef1"). InnerVolumeSpecName "kube-api-access-m6cbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.209895 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7l7r\" (UniqueName: \"kubernetes.io/projected/a372a06b-1453-4141-b82b-96af30ceb608-kube-api-access-k7l7r\") pod \"a372a06b-1453-4141-b82b-96af30ceb608\" (UID: \"a372a06b-1453-4141-b82b-96af30ceb608\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.210047 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57dc900-6506-4a69-b49a-3bd5dd6a88db-operator-scripts\") pod \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\" (UID: \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.210169 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a372a06b-1453-4141-b82b-96af30ceb608-operator-scripts\") pod \"a372a06b-1453-4141-b82b-96af30ceb608\" (UID: \"a372a06b-1453-4141-b82b-96af30ceb608\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.210252 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc6mt\" (UniqueName: \"kubernetes.io/projected/a57dc900-6506-4a69-b49a-3bd5dd6a88db-kube-api-access-lc6mt\") pod \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\" (UID: \"a57dc900-6506-4a69-b49a-3bd5dd6a88db\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.210742 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6cbs\" (UniqueName: \"kubernetes.io/projected/44b3ba14-81e9-4338-bc3e-297128ea5ef1-kube-api-access-m6cbs\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.210758 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhjqt\" (UniqueName: \"kubernetes.io/projected/795cf078-881c-4373-9e3c-13afd8662252-kube-api-access-bhjqt\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.211693 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a57dc900-6506-4a69-b49a-3bd5dd6a88db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a57dc900-6506-4a69-b49a-3bd5dd6a88db" (UID: "a57dc900-6506-4a69-b49a-3bd5dd6a88db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.211787 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a372a06b-1453-4141-b82b-96af30ceb608-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a372a06b-1453-4141-b82b-96af30ceb608" (UID: "a372a06b-1453-4141-b82b-96af30ceb608"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.214206 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a57dc900-6506-4a69-b49a-3bd5dd6a88db-kube-api-access-lc6mt" (OuterVolumeSpecName: "kube-api-access-lc6mt") pod "a57dc900-6506-4a69-b49a-3bd5dd6a88db" (UID: "a57dc900-6506-4a69-b49a-3bd5dd6a88db"). InnerVolumeSpecName "kube-api-access-lc6mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.221591 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a372a06b-1453-4141-b82b-96af30ceb608-kube-api-access-k7l7r" (OuterVolumeSpecName: "kube-api-access-k7l7r") pod "a372a06b-1453-4141-b82b-96af30ceb608" (UID: "a372a06b-1453-4141-b82b-96af30ceb608"). InnerVolumeSpecName "kube-api-access-k7l7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.301966 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.316375 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a57dc900-6506-4a69-b49a-3bd5dd6a88db-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.316438 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a372a06b-1453-4141-b82b-96af30ceb608-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.316473 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc6mt\" (UniqueName: \"kubernetes.io/projected/a57dc900-6506-4a69-b49a-3bd5dd6a88db-kube-api-access-lc6mt\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.316485 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7l7r\" (UniqueName: \"kubernetes.io/projected/a372a06b-1453-4141-b82b-96af30ceb608-kube-api-access-k7l7r\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.336809 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-79c64" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.353955 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.404630 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.417299 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afdb36a9-143a-4bbc-a826-cdfa078e5d68-operator-scripts\") pod \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\" (UID: \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.417461 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt6sq\" (UniqueName: \"kubernetes.io/projected/afdb36a9-143a-4bbc-a826-cdfa078e5d68-kube-api-access-kt6sq\") pod \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\" (UID: \"afdb36a9-143a-4bbc-a826-cdfa078e5d68\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.425231 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afdb36a9-143a-4bbc-a826-cdfa078e5d68-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "afdb36a9-143a-4bbc-a826-cdfa078e5d68" (UID: "afdb36a9-143a-4bbc-a826-cdfa078e5d68"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.435150 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdb36a9-143a-4bbc-a826-cdfa078e5d68-kube-api-access-kt6sq" (OuterVolumeSpecName: "kube-api-access-kt6sq") pod "afdb36a9-143a-4bbc-a826-cdfa078e5d68" (UID: "afdb36a9-143a-4bbc-a826-cdfa078e5d68"). InnerVolumeSpecName "kube-api-access-kt6sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.519089 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxvcm\" (UniqueName: \"kubernetes.io/projected/5df8f82c-aba7-4d44-b4bd-c3a616afc480-kube-api-access-rxvcm\") pod \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\" (UID: \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.519307 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8jmk\" (UniqueName: \"kubernetes.io/projected/16de9a20-7669-4935-9995-a5d1172e68e3-kube-api-access-l8jmk\") pod \"16de9a20-7669-4935-9995-a5d1172e68e3\" (UID: \"16de9a20-7669-4935-9995-a5d1172e68e3\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.519343 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41cf511d-549f-4b68-bb95-40a73cddbf83-operator-scripts\") pod \"41cf511d-549f-4b68-bb95-40a73cddbf83\" (UID: \"41cf511d-549f-4b68-bb95-40a73cddbf83\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.519391 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbnbf\" (UniqueName: \"kubernetes.io/projected/41cf511d-549f-4b68-bb95-40a73cddbf83-kube-api-access-cbnbf\") pod \"41cf511d-549f-4b68-bb95-40a73cddbf83\" (UID: \"41cf511d-549f-4b68-bb95-40a73cddbf83\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.519531 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5df8f82c-aba7-4d44-b4bd-c3a616afc480-operator-scripts\") pod \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\" (UID: \"5df8f82c-aba7-4d44-b4bd-c3a616afc480\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.519659 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16de9a20-7669-4935-9995-a5d1172e68e3-operator-scripts\") pod \"16de9a20-7669-4935-9995-a5d1172e68e3\" (UID: \"16de9a20-7669-4935-9995-a5d1172e68e3\") " Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.520270 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41cf511d-549f-4b68-bb95-40a73cddbf83-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "41cf511d-549f-4b68-bb95-40a73cddbf83" (UID: "41cf511d-549f-4b68-bb95-40a73cddbf83"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.520838 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16de9a20-7669-4935-9995-a5d1172e68e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "16de9a20-7669-4935-9995-a5d1172e68e3" (UID: "16de9a20-7669-4935-9995-a5d1172e68e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.520904 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5df8f82c-aba7-4d44-b4bd-c3a616afc480-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5df8f82c-aba7-4d44-b4bd-c3a616afc480" (UID: "5df8f82c-aba7-4d44-b4bd-c3a616afc480"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.521073 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/afdb36a9-143a-4bbc-a826-cdfa078e5d68-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.521132 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/41cf511d-549f-4b68-bb95-40a73cddbf83-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.521146 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt6sq\" (UniqueName: \"kubernetes.io/projected/afdb36a9-143a-4bbc-a826-cdfa078e5d68-kube-api-access-kt6sq\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.522831 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16de9a20-7669-4935-9995-a5d1172e68e3-kube-api-access-l8jmk" (OuterVolumeSpecName: "kube-api-access-l8jmk") pod "16de9a20-7669-4935-9995-a5d1172e68e3" (UID: "16de9a20-7669-4935-9995-a5d1172e68e3"). InnerVolumeSpecName "kube-api-access-l8jmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.522890 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5df8f82c-aba7-4d44-b4bd-c3a616afc480-kube-api-access-rxvcm" (OuterVolumeSpecName: "kube-api-access-rxvcm") pod "5df8f82c-aba7-4d44-b4bd-c3a616afc480" (UID: "5df8f82c-aba7-4d44-b4bd-c3a616afc480"). InnerVolumeSpecName "kube-api-access-rxvcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.526441 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41cf511d-549f-4b68-bb95-40a73cddbf83-kube-api-access-cbnbf" (OuterVolumeSpecName: "kube-api-access-cbnbf") pod "41cf511d-549f-4b68-bb95-40a73cddbf83" (UID: "41cf511d-549f-4b68-bb95-40a73cddbf83"). InnerVolumeSpecName "kube-api-access-cbnbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.629944 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5df8f82c-aba7-4d44-b4bd-c3a616afc480-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.629978 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/16de9a20-7669-4935-9995-a5d1172e68e3-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.629991 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxvcm\" (UniqueName: \"kubernetes.io/projected/5df8f82c-aba7-4d44-b4bd-c3a616afc480-kube-api-access-rxvcm\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.630006 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8jmk\" (UniqueName: \"kubernetes.io/projected/16de9a20-7669-4935-9995-a5d1172e68e3-kube-api-access-l8jmk\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.630018 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbnbf\" (UniqueName: \"kubernetes.io/projected/41cf511d-549f-4b68-bb95-40a73cddbf83-kube-api-access-cbnbf\") on node \"crc\" DevicePath \"\"" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.927698 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-68rr4" event={"ID":"cb6ad61a-8190-4d4d-987c-f609c1e8cf44","Type":"ContainerStarted","Data":"2ab20d9a010b2164f31b5440bef346b4bb35d8a31cc6ba96c6bfba16eaf5f6b4"} Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.934034 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"5428edb3cba79503dbea53b80df68874d85edf93d9d07808d973b19ce0b44e74"} Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.934085 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"b6f998a9aa074b3620df4745192c26fc44e1dac816757fdafe9c6b3ce36a9910"} Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.934098 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"3a9d2d6585c2611c7202f6d141f7f1e94f8c08fef0318d0939d793ccbbfb6619"} Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.934109 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"a1e90715aff297128ba7299afb9db5f70a7f87b3a062d2777ffff14e559ee851"} Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.935354 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-gwmvk" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.936150 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f033-account-create-update-w89dh" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.936488 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-s8j7d" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.936527 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jbwmt" event={"ID":"66f0fe7a-9061-4bc2-a957-f9d2926401d7","Type":"ContainerStarted","Data":"ef729b5a2939b5b0e257b95388618106527be7a6dc2ddda746a607596abf9c7e"} Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.936561 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-6252-account-create-update-cz7bz" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.937090 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-67ffc" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.937131 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-f5a5-account-create-update-jkh64" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.937183 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6b41-account-create-update-9qd64" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.937844 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-79c64" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.969517 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-68rr4" podStartSLOduration=2.778043525 podStartE2EDuration="39.969494837s" podCreationTimestamp="2026-01-06 14:57:16 +0000 UTC" firstStartedPulling="2026-01-06 14:57:17.881311832 +0000 UTC m=+1234.508778150" lastFinishedPulling="2026-01-06 14:57:55.072763144 +0000 UTC m=+1271.700229462" observedRunningTime="2026-01-06 14:57:55.947170307 +0000 UTC m=+1272.574636625" watchObservedRunningTime="2026-01-06 14:57:55.969494837 +0000 UTC m=+1272.596961155" Jan 06 14:57:55 crc kubenswrapper[4744]: I0106 14:57:55.981657 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-jbwmt" podStartSLOduration=3.256304706 podStartE2EDuration="14.981633978s" podCreationTimestamp="2026-01-06 14:57:41 +0000 UTC" firstStartedPulling="2026-01-06 14:57:43.34771446 +0000 UTC m=+1259.975180778" lastFinishedPulling="2026-01-06 14:57:55.073043732 +0000 UTC m=+1271.700510050" observedRunningTime="2026-01-06 14:57:55.963283493 +0000 UTC m=+1272.590749831" watchObservedRunningTime="2026-01-06 14:57:55.981633978 +0000 UTC m=+1272.609100296" Jan 06 14:57:56 crc kubenswrapper[4744]: I0106 14:57:56.986259 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c603cde3-b7b0-4851-99df-2abb0aa6b1c4","Type":"ContainerStarted","Data":"e3f064d63ec014118e3db541b81c893be34230401531100bee1772f737a95cc9"} Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.027835 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=44.71235296 podStartE2EDuration="1m14.027799254s" podCreationTimestamp="2026-01-06 14:56:43 +0000 UTC" firstStartedPulling="2026-01-06 14:57:17.468355672 +0000 UTC m=+1234.095821990" lastFinishedPulling="2026-01-06 14:57:46.783801966 +0000 UTC m=+1263.411268284" observedRunningTime="2026-01-06 14:57:57.018453187 +0000 UTC m=+1273.645919515" watchObservedRunningTime="2026-01-06 14:57:57.027799254 +0000 UTC m=+1273.655265572" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.362279 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-f6xxt"] Jan 06 14:57:57 crc kubenswrapper[4744]: E0106 14:57:57.362944 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b3ba14-81e9-4338-bc3e-297128ea5ef1" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.362964 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b3ba14-81e9-4338-bc3e-297128ea5ef1" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: E0106 14:57:57.362981 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41cf511d-549f-4b68-bb95-40a73cddbf83" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.362987 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="41cf511d-549f-4b68-bb95-40a73cddbf83" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: E0106 14:57:57.363008 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16de9a20-7669-4935-9995-a5d1172e68e3" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363018 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="16de9a20-7669-4935-9995-a5d1172e68e3" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: E0106 14:57:57.363035 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a57dc900-6506-4a69-b49a-3bd5dd6a88db" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363041 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a57dc900-6506-4a69-b49a-3bd5dd6a88db" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: E0106 14:57:57.363047 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795cf078-881c-4373-9e3c-13afd8662252" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363054 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="795cf078-881c-4373-9e3c-13afd8662252" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: E0106 14:57:57.363067 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdb36a9-143a-4bbc-a826-cdfa078e5d68" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363072 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdb36a9-143a-4bbc-a826-cdfa078e5d68" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: E0106 14:57:57.363085 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5df8f82c-aba7-4d44-b4bd-c3a616afc480" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363091 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5df8f82c-aba7-4d44-b4bd-c3a616afc480" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: E0106 14:57:57.363101 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a372a06b-1453-4141-b82b-96af30ceb608" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363107 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a372a06b-1453-4141-b82b-96af30ceb608" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363538 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="16de9a20-7669-4935-9995-a5d1172e68e3" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363562 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="41cf511d-549f-4b68-bb95-40a73cddbf83" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363581 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdb36a9-143a-4bbc-a826-cdfa078e5d68" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363591 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a57dc900-6506-4a69-b49a-3bd5dd6a88db" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363602 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a372a06b-1453-4141-b82b-96af30ceb608" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363610 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5df8f82c-aba7-4d44-b4bd-c3a616afc480" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363621 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b3ba14-81e9-4338-bc3e-297128ea5ef1" containerName="mariadb-account-create-update" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.363630 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="795cf078-881c-4373-9e3c-13afd8662252" containerName="mariadb-database-create" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.364678 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.367465 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.388349 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-f6xxt"] Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.480048 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n7kq\" (UniqueName: \"kubernetes.io/projected/def77701-47a7-4a7d-befc-75f5352554b6-kube-api-access-7n7kq\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.480089 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.480145 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.480245 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.480312 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.480342 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-config\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.582508 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.582607 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.582715 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.582785 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.582816 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-config\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.582906 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n7kq\" (UniqueName: \"kubernetes.io/projected/def77701-47a7-4a7d-befc-75f5352554b6-kube-api-access-7n7kq\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.584084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.584731 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.585031 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.585833 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.586062 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-config\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.607800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n7kq\" (UniqueName: \"kubernetes.io/projected/def77701-47a7-4a7d-befc-75f5352554b6-kube-api-access-7n7kq\") pod \"dnsmasq-dns-6d5b6d6b67-f6xxt\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:57 crc kubenswrapper[4744]: I0106 14:57:57.683559 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:57:58 crc kubenswrapper[4744]: I0106 14:57:58.276152 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-f6xxt"] Jan 06 14:57:59 crc kubenswrapper[4744]: I0106 14:57:59.007987 4744 generic.go:334] "Generic (PLEG): container finished" podID="0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf" containerID="e98b63d3ec96f756dcbed447564b013bcd722583963ce64dddb6a2fe9221dbfa" exitCode=0 Jan 06 14:57:59 crc kubenswrapper[4744]: I0106 14:57:59.008089 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf","Type":"ContainerDied","Data":"e98b63d3ec96f756dcbed447564b013bcd722583963ce64dddb6a2fe9221dbfa"} Jan 06 14:57:59 crc kubenswrapper[4744]: I0106 14:57:59.009663 4744 generic.go:334] "Generic (PLEG): container finished" podID="def77701-47a7-4a7d-befc-75f5352554b6" containerID="87e558bdac3b358a19862e5588c16cf2df9999b4b7ff455ac1e4e3d649928669" exitCode=0 Jan 06 14:57:59 crc kubenswrapper[4744]: I0106 14:57:59.009683 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" event={"ID":"def77701-47a7-4a7d-befc-75f5352554b6","Type":"ContainerDied","Data":"87e558bdac3b358a19862e5588c16cf2df9999b4b7ff455ac1e4e3d649928669"} Jan 06 14:57:59 crc kubenswrapper[4744]: I0106 14:57:59.009708 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" event={"ID":"def77701-47a7-4a7d-befc-75f5352554b6","Type":"ContainerStarted","Data":"92a7d5c0e51adeea5d167f482368e090b261027c8e56936f2f584815f311d820"} Jan 06 14:58:00 crc kubenswrapper[4744]: I0106 14:58:00.031587 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf","Type":"ContainerStarted","Data":"c7c79e84b6c96fa11af53692a7fe82725e8cfe5652f1f65dff8c7d8ca7c35f03"} Jan 06 14:58:00 crc kubenswrapper[4744]: I0106 14:58:00.034612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" event={"ID":"def77701-47a7-4a7d-befc-75f5352554b6","Type":"ContainerStarted","Data":"aa6faeaefb8d50626c37fa1461f08d96b9b5da11700a654c712adad5b879a3e0"} Jan 06 14:58:00 crc kubenswrapper[4744]: I0106 14:58:00.034970 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:58:00 crc kubenswrapper[4744]: I0106 14:58:00.077228 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" podStartSLOduration=3.077195104 podStartE2EDuration="3.077195104s" podCreationTimestamp="2026-01-06 14:57:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:58:00.062483035 +0000 UTC m=+1276.689949353" watchObservedRunningTime="2026-01-06 14:58:00.077195104 +0000 UTC m=+1276.704661422" Jan 06 14:58:05 crc kubenswrapper[4744]: I0106 14:58:05.098064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf","Type":"ContainerStarted","Data":"bcd7d429c50dfc61c2e59b903820c7ceb05f35711505ef5ad82ccb3ad2f1d6c0"} Jan 06 14:58:05 crc kubenswrapper[4744]: I0106 14:58:05.098699 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf","Type":"ContainerStarted","Data":"747ad65cd0b0d354aae6dc8eaaffcd443ce4a0590ba9ef4d0ac741a753eec20b"} Jan 06 14:58:05 crc kubenswrapper[4744]: I0106 14:58:05.132041 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=39.132022727 podStartE2EDuration="39.132022727s" podCreationTimestamp="2026-01-06 14:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:58:05.124743374 +0000 UTC m=+1281.752209732" watchObservedRunningTime="2026-01-06 14:58:05.132022727 +0000 UTC m=+1281.759489055" Jan 06 14:58:06 crc kubenswrapper[4744]: I0106 14:58:06.720109 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Jan 06 14:58:07 crc kubenswrapper[4744]: I0106 14:58:07.686343 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:58:07 crc kubenswrapper[4744]: I0106 14:58:07.791613 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9thnp"] Jan 06 14:58:07 crc kubenswrapper[4744]: I0106 14:58:07.791914 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" podUID="dd70b136-b28e-4bb2-b85f-e0718fb3ad67" containerName="dnsmasq-dns" containerID="cri-o://2da3e48201e6ebeb7f81d43329815c196b5f93a2c494e874c63ee3331711065d" gracePeriod=10 Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.137037 4744 generic.go:334] "Generic (PLEG): container finished" podID="dd70b136-b28e-4bb2-b85f-e0718fb3ad67" containerID="2da3e48201e6ebeb7f81d43329815c196b5f93a2c494e874c63ee3331711065d" exitCode=0 Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.137104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" event={"ID":"dd70b136-b28e-4bb2-b85f-e0718fb3ad67","Type":"ContainerDied","Data":"2da3e48201e6ebeb7f81d43329815c196b5f93a2c494e874c63ee3331711065d"} Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.430420 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.458030 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-nb\") pod \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.458365 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-sb\") pod \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.458488 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgwq4\" (UniqueName: \"kubernetes.io/projected/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-kube-api-access-bgwq4\") pod \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.458657 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-config\") pod \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.458738 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-dns-svc\") pod \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\" (UID: \"dd70b136-b28e-4bb2-b85f-e0718fb3ad67\") " Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.467425 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-kube-api-access-bgwq4" (OuterVolumeSpecName: "kube-api-access-bgwq4") pod "dd70b136-b28e-4bb2-b85f-e0718fb3ad67" (UID: "dd70b136-b28e-4bb2-b85f-e0718fb3ad67"). InnerVolumeSpecName "kube-api-access-bgwq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.521787 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd70b136-b28e-4bb2-b85f-e0718fb3ad67" (UID: "dd70b136-b28e-4bb2-b85f-e0718fb3ad67"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.534075 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-config" (OuterVolumeSpecName: "config") pod "dd70b136-b28e-4bb2-b85f-e0718fb3ad67" (UID: "dd70b136-b28e-4bb2-b85f-e0718fb3ad67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.539306 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dd70b136-b28e-4bb2-b85f-e0718fb3ad67" (UID: "dd70b136-b28e-4bb2-b85f-e0718fb3ad67"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.543907 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dd70b136-b28e-4bb2-b85f-e0718fb3ad67" (UID: "dd70b136-b28e-4bb2-b85f-e0718fb3ad67"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.563608 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.563641 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.563651 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.563659 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:09 crc kubenswrapper[4744]: I0106 14:58:09.563668 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgwq4\" (UniqueName: \"kubernetes.io/projected/dd70b136-b28e-4bb2-b85f-e0718fb3ad67-kube-api-access-bgwq4\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:10 crc kubenswrapper[4744]: I0106 14:58:10.156762 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" event={"ID":"dd70b136-b28e-4bb2-b85f-e0718fb3ad67","Type":"ContainerDied","Data":"1c45e81f00c94715f0cc1650785e5318c837da0b1bed7b2e87a46f483f6231a0"} Jan 06 14:58:10 crc kubenswrapper[4744]: I0106 14:58:10.156815 4744 scope.go:117] "RemoveContainer" containerID="2da3e48201e6ebeb7f81d43329815c196b5f93a2c494e874c63ee3331711065d" Jan 06 14:58:10 crc kubenswrapper[4744]: I0106 14:58:10.156958 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-9thnp" Jan 06 14:58:10 crc kubenswrapper[4744]: I0106 14:58:10.197256 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9thnp"] Jan 06 14:58:10 crc kubenswrapper[4744]: I0106 14:58:10.208923 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-9thnp"] Jan 06 14:58:10 crc kubenswrapper[4744]: I0106 14:58:10.216091 4744 scope.go:117] "RemoveContainer" containerID="b8e76c4d53d93ae810ab2620fcc4022d458b91aba0cb92e5f7264fbbb909b39b" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.064506 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pbg7p"] Jan 06 14:58:11 crc kubenswrapper[4744]: E0106 14:58:11.065037 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd70b136-b28e-4bb2-b85f-e0718fb3ad67" containerName="init" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.065061 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd70b136-b28e-4bb2-b85f-e0718fb3ad67" containerName="init" Jan 06 14:58:11 crc kubenswrapper[4744]: E0106 14:58:11.065114 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd70b136-b28e-4bb2-b85f-e0718fb3ad67" containerName="dnsmasq-dns" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.065122 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd70b136-b28e-4bb2-b85f-e0718fb3ad67" containerName="dnsmasq-dns" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.073034 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd70b136-b28e-4bb2-b85f-e0718fb3ad67" containerName="dnsmasq-dns" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.074919 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.097137 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pbg7p"] Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.166624 4744 generic.go:334] "Generic (PLEG): container finished" podID="66f0fe7a-9061-4bc2-a957-f9d2926401d7" containerID="ef729b5a2939b5b0e257b95388618106527be7a6dc2ddda746a607596abf9c7e" exitCode=0 Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.166673 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jbwmt" event={"ID":"66f0fe7a-9061-4bc2-a957-f9d2926401d7","Type":"ContainerDied","Data":"ef729b5a2939b5b0e257b95388618106527be7a6dc2ddda746a607596abf9c7e"} Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.202277 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-catalog-content\") pod \"redhat-operators-pbg7p\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.202326 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-utilities\") pod \"redhat-operators-pbg7p\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.202397 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmfcw\" (UniqueName: \"kubernetes.io/projected/e584b5b6-874e-48ea-b531-267a2c20562b-kube-api-access-hmfcw\") pod \"redhat-operators-pbg7p\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.304256 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-catalog-content\") pod \"redhat-operators-pbg7p\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.304323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-utilities\") pod \"redhat-operators-pbg7p\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.304483 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmfcw\" (UniqueName: \"kubernetes.io/projected/e584b5b6-874e-48ea-b531-267a2c20562b-kube-api-access-hmfcw\") pod \"redhat-operators-pbg7p\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.304963 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-catalog-content\") pod \"redhat-operators-pbg7p\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.305209 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-utilities\") pod \"redhat-operators-pbg7p\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.324897 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmfcw\" (UniqueName: \"kubernetes.io/projected/e584b5b6-874e-48ea-b531-267a2c20562b-kube-api-access-hmfcw\") pod \"redhat-operators-pbg7p\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.406298 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.727445 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd70b136-b28e-4bb2-b85f-e0718fb3ad67" path="/var/lib/kubelet/pods/dd70b136-b28e-4bb2-b85f-e0718fb3ad67/volumes" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.728151 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.729866 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Jan 06 14:58:11 crc kubenswrapper[4744]: I0106 14:58:11.881902 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pbg7p"] Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.178745 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbg7p" event={"ID":"e584b5b6-874e-48ea-b531-267a2c20562b","Type":"ContainerStarted","Data":"07b6c3bd77195704a22a462b8b1e61103d0ebf4b80c8cd98eb073df55f197301"} Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.184358 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.675741 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.780989 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mx8dp\" (UniqueName: \"kubernetes.io/projected/66f0fe7a-9061-4bc2-a957-f9d2926401d7-kube-api-access-mx8dp\") pod \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.781113 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-combined-ca-bundle\") pod \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.781365 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-config-data\") pod \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\" (UID: \"66f0fe7a-9061-4bc2-a957-f9d2926401d7\") " Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.787345 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66f0fe7a-9061-4bc2-a957-f9d2926401d7-kube-api-access-mx8dp" (OuterVolumeSpecName: "kube-api-access-mx8dp") pod "66f0fe7a-9061-4bc2-a957-f9d2926401d7" (UID: "66f0fe7a-9061-4bc2-a957-f9d2926401d7"). InnerVolumeSpecName "kube-api-access-mx8dp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.831785 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66f0fe7a-9061-4bc2-a957-f9d2926401d7" (UID: "66f0fe7a-9061-4bc2-a957-f9d2926401d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.843914 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-config-data" (OuterVolumeSpecName: "config-data") pod "66f0fe7a-9061-4bc2-a957-f9d2926401d7" (UID: "66f0fe7a-9061-4bc2-a957-f9d2926401d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.884306 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mx8dp\" (UniqueName: \"kubernetes.io/projected/66f0fe7a-9061-4bc2-a957-f9d2926401d7-kube-api-access-mx8dp\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.884339 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:12 crc kubenswrapper[4744]: I0106 14:58:12.884350 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66f0fe7a-9061-4bc2-a957-f9d2926401d7-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.190480 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-jbwmt" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.190825 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-jbwmt" event={"ID":"66f0fe7a-9061-4bc2-a957-f9d2926401d7","Type":"ContainerDied","Data":"58e753163e29654a79be41d8f1635fbc872d800b8b3743fff7e6db8e7cf4f18d"} Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.190864 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58e753163e29654a79be41d8f1635fbc872d800b8b3743fff7e6db8e7cf4f18d" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.194404 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbg7p" event={"ID":"e584b5b6-874e-48ea-b531-267a2c20562b","Type":"ContainerStarted","Data":"7b358706df1274187d1b41ce1016e8642b2ae2f7ec73a034eb6f9386e8c12f3f"} Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.415751 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-7xxr2"] Jan 06 14:58:13 crc kubenswrapper[4744]: E0106 14:58:13.416471 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f0fe7a-9061-4bc2-a957-f9d2926401d7" containerName="keystone-db-sync" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.416500 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f0fe7a-9061-4bc2-a957-f9d2926401d7" containerName="keystone-db-sync" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.416869 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="66f0fe7a-9061-4bc2-a957-f9d2926401d7" containerName="keystone-db-sync" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.443274 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.445963 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-7xxr2"] Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.498567 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.498635 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-config\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.498726 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.498784 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.498831 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.498853 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbzhw\" (UniqueName: \"kubernetes.io/projected/52611203-15d5-41ba-b286-06d7058521d5-kube-api-access-rbzhw\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.537397 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dkghr"] Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.539757 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.544802 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.545013 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.547577 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.549267 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.553935 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8hqr4" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.566323 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dkghr"] Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.608548 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-config-data\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.608613 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.608679 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-credential-keys\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.608750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.608808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w8xs\" (UniqueName: \"kubernetes.io/projected/d6889ed3-f69b-428a-a8f7-68e5ff3550df-kube-api-access-6w8xs\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.608880 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.608905 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-fernet-keys\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.608940 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbzhw\" (UniqueName: \"kubernetes.io/projected/52611203-15d5-41ba-b286-06d7058521d5-kube-api-access-rbzhw\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.609056 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-scripts\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.609123 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-combined-ca-bundle\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.609247 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.609285 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-config\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.609540 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.609968 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.610108 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-config\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.610372 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.610719 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.638991 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-6ltjv"] Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.640426 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.666025 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbzhw\" (UniqueName: \"kubernetes.io/projected/52611203-15d5-41ba-b286-06d7058521d5-kube-api-access-rbzhw\") pod \"dnsmasq-dns-6f8c45789f-7xxr2\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.668176 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-zl99k" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.668269 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.706135 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-6ltjv"] Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.710685 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-combined-ca-bundle\") pod \"heat-db-sync-6ltjv\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.710877 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-config-data\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.710990 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-credential-keys\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.711165 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w8xs\" (UniqueName: \"kubernetes.io/projected/d6889ed3-f69b-428a-a8f7-68e5ff3550df-kube-api-access-6w8xs\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.711272 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62p9x\" (UniqueName: \"kubernetes.io/projected/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-kube-api-access-62p9x\") pod \"heat-db-sync-6ltjv\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.711356 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-fernet-keys\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.711483 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-scripts\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.711566 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-combined-ca-bundle\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.711654 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-config-data\") pod \"heat-db-sync-6ltjv\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.716218 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-credential-keys\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.717351 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-fernet-keys\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.717582 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-config-data\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.726462 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-combined-ca-bundle\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.736017 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w8xs\" (UniqueName: \"kubernetes.io/projected/d6889ed3-f69b-428a-a8f7-68e5ff3550df-kube-api-access-6w8xs\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.737245 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-scripts\") pod \"keystone-bootstrap-dkghr\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.815911 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.816521 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-config-data\") pod \"heat-db-sync-6ltjv\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.816583 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-combined-ca-bundle\") pod \"heat-db-sync-6ltjv\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.816697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62p9x\" (UniqueName: \"kubernetes.io/projected/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-kube-api-access-62p9x\") pod \"heat-db-sync-6ltjv\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.839071 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-config-data\") pod \"heat-db-sync-6ltjv\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.850566 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-combined-ca-bundle\") pod \"heat-db-sync-6ltjv\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.851956 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62p9x\" (UniqueName: \"kubernetes.io/projected/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-kube-api-access-62p9x\") pod \"heat-db-sync-6ltjv\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.864942 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.901679 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6ltjv" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.940024 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-vdnt5"] Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.943648 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.946852 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.947097 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 06 14:58:13 crc kubenswrapper[4744]: I0106 14:58:13.947137 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7hzlw" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.001341 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vdnt5"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.055781 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-db-sync-config-data\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.055885 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-scripts\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.055928 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbzxc\" (UniqueName: \"kubernetes.io/projected/2b74efd3-9c21-4555-bfa8-5c1652b380dc-kube-api-access-bbzxc\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.055985 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-config-data\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.056173 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b74efd3-9c21-4555-bfa8-5c1652b380dc-etc-machine-id\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.056551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-combined-ca-bundle\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.087917 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-m88z9"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.095908 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.110724 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-vq66f" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.110994 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.115421 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.191354 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-combined-ca-bundle\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.191761 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-db-sync-config-data\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.201908 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-m88z9"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.211404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-scripts\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.211446 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbzxc\" (UniqueName: \"kubernetes.io/projected/2b74efd3-9c21-4555-bfa8-5c1652b380dc-kube-api-access-bbzxc\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.211497 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-config-data\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.211634 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b74efd3-9c21-4555-bfa8-5c1652b380dc-etc-machine-id\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.211994 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b74efd3-9c21-4555-bfa8-5c1652b380dc-etc-machine-id\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.216056 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-m8c54"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.217523 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.219118 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-config-data\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.230970 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-combined-ca-bundle\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.235912 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-scripts\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.241700 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-7xxr2"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.245689 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kzz6d" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.246021 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.254330 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-db-sync-config-data\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.262225 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-m8c54"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.263556 4744 generic.go:334] "Generic (PLEG): container finished" podID="e584b5b6-874e-48ea-b531-267a2c20562b" containerID="7b358706df1274187d1b41ce1016e8642b2ae2f7ec73a034eb6f9386e8c12f3f" exitCode=0 Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.263612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbg7p" event={"ID":"e584b5b6-874e-48ea-b531-267a2c20562b","Type":"ContainerDied","Data":"7b358706df1274187d1b41ce1016e8642b2ae2f7ec73a034eb6f9386e8c12f3f"} Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.289014 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbzxc\" (UniqueName: \"kubernetes.io/projected/2b74efd3-9c21-4555-bfa8-5c1652b380dc-kube-api-access-bbzxc\") pod \"cinder-db-sync-vdnt5\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.292048 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-xm2rp"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.295345 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.308826 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-dfxz9"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.310264 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.313650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-config\") pod \"neutron-db-sync-m88z9\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.313852 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gbb6\" (UniqueName: \"kubernetes.io/projected/4d50465b-09ea-4ff4-aa47-12f1c49ff207-kube-api-access-7gbb6\") pod \"barbican-db-sync-m8c54\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.313904 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-combined-ca-bundle\") pod \"barbican-db-sync-m8c54\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.313937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqtcx\" (UniqueName: \"kubernetes.io/projected/d9e793d1-65f3-43a8-914b-112690afc33f-kube-api-access-zqtcx\") pod \"neutron-db-sync-m88z9\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.313958 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-db-sync-config-data\") pod \"barbican-db-sync-m8c54\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.313986 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-combined-ca-bundle\") pod \"neutron-db-sync-m88z9\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.334758 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.334985 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-m65j6" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.342802 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vdnt5" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.352416 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-xm2rp"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.352540 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.375297 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dfxz9"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.403257 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.415531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.415697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gbb6\" (UniqueName: \"kubernetes.io/projected/4d50465b-09ea-4ff4-aa47-12f1c49ff207-kube-api-access-7gbb6\") pod \"barbican-db-sync-m8c54\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.415796 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-combined-ca-bundle\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.415887 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-combined-ca-bundle\") pod \"barbican-db-sync-m8c54\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.415978 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqtcx\" (UniqueName: \"kubernetes.io/projected/d9e793d1-65f3-43a8-914b-112690afc33f-kube-api-access-zqtcx\") pod \"neutron-db-sync-m88z9\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.416099 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-db-sync-config-data\") pod \"barbican-db-sync-m8c54\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.416207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.416293 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-combined-ca-bundle\") pod \"neutron-db-sync-m88z9\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.416388 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-kube-api-access-5n6zx\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.416550 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.416633 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.416754 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-config-data\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.416893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-config\") pod \"neutron-db-sync-m88z9\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.417068 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m65r\" (UniqueName: \"kubernetes.io/projected/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-kube-api-access-9m65r\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.417198 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-config\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.417332 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-scripts\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.417471 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-logs\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.423139 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-combined-ca-bundle\") pod \"barbican-db-sync-m8c54\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.426818 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-db-sync-config-data\") pod \"barbican-db-sync-m8c54\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.427270 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.427320 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.444521 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.448967 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.450359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-config\") pod \"neutron-db-sync-m88z9\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.461882 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-combined-ca-bundle\") pod \"neutron-db-sync-m88z9\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.467830 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.468066 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.473476 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8cc1261767d494e24246d9d799ed9782619069fb6a47b29f76fd6dc783322bbb"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.473550 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://8cc1261767d494e24246d9d799ed9782619069fb6a47b29f76fd6dc783322bbb" gracePeriod=600 Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.489965 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gbb6\" (UniqueName: \"kubernetes.io/projected/4d50465b-09ea-4ff4-aa47-12f1c49ff207-kube-api-access-7gbb6\") pod \"barbican-db-sync-m8c54\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.491836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqtcx\" (UniqueName: \"kubernetes.io/projected/d9e793d1-65f3-43a8-914b-112690afc33f-kube-api-access-zqtcx\") pod \"neutron-db-sync-m88z9\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.492571 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m88z9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.529566 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-config-data\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.529668 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.529738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-run-httpd\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.529765 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-log-httpd\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.529816 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m65r\" (UniqueName: \"kubernetes.io/projected/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-kube-api-access-9m65r\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.529851 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-config\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.529878 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-scripts\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.529907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-scripts\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.529956 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-logs\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.530008 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.530041 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.530062 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-combined-ca-bundle\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.530101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.530131 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-kube-api-access-5n6zx\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.530207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktcb5\" (UniqueName: \"kubernetes.io/projected/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-kube-api-access-ktcb5\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.530234 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-config-data\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.530301 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.530324 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.531553 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.532056 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.532493 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-logs\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.533068 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.536284 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-config\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.538566 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.539516 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-combined-ca-bundle\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.578215 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-config-data\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.579278 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.581449 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m65r\" (UniqueName: \"kubernetes.io/projected/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-kube-api-access-9m65r\") pod \"dnsmasq-dns-fcfdd6f9f-xm2rp\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.589973 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-scripts\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.592375 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-kube-api-access-5n6zx\") pod \"placement-db-sync-dfxz9\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.610063 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-m8c54" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.651322 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktcb5\" (UniqueName: \"kubernetes.io/projected/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-kube-api-access-ktcb5\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.651632 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-config-data\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.651930 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.652025 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-run-httpd\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.652101 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-log-httpd\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.652269 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-scripts\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.652512 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.655411 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-run-httpd\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.656659 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.657029 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-log-httpd\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.657070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-config-data\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.660240 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-scripts\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.674212 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.685730 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.690564 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktcb5\" (UniqueName: \"kubernetes.io/projected/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-kube-api-access-ktcb5\") pod \"ceilometer-0\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.752032 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dfxz9" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.856074 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 14:58:14 crc kubenswrapper[4744]: I0106 14:58:14.870998 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-7xxr2"] Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.108455 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-6ltjv"] Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.281633 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6ltjv" event={"ID":"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b","Type":"ContainerStarted","Data":"99bcb3b0f6adef94a66559e37041aec8a53c8c8dcf357f45ed22d76e6f7be89f"} Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.307637 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="8cc1261767d494e24246d9d799ed9782619069fb6a47b29f76fd6dc783322bbb" exitCode=0 Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.307684 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"8cc1261767d494e24246d9d799ed9782619069fb6a47b29f76fd6dc783322bbb"} Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.307740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c"} Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.307759 4744 scope.go:117] "RemoveContainer" containerID="7b95bf86d9d16dc50821869dcd40536e357bac00adf9429e476ea048ffe56203" Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.309889 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" event={"ID":"52611203-15d5-41ba-b286-06d7058521d5","Type":"ContainerStarted","Data":"d044b772dc9ef968f2198e1177ce108d55f127731df181a7ce9e01738e6afb95"} Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.348143 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-vdnt5"] Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.517956 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dkghr"] Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.531432 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-m88z9"] Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.822841 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-xm2rp"] Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.827209 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-m8c54"] Jan 06 14:58:15 crc kubenswrapper[4744]: W0106 14:58:15.893589 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0832cc9a_5d34_4e11_8d3b_cc7c0674b73a.slice/crio-f1ecd45a43453fcf957dc964d93a7f7d299de053fc5b0a834c1733630a2f3e1f WatchSource:0}: Error finding container f1ecd45a43453fcf957dc964d93a7f7d299de053fc5b0a834c1733630a2f3e1f: Status 404 returned error can't find the container with id f1ecd45a43453fcf957dc964d93a7f7d299de053fc5b0a834c1733630a2f3e1f Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.938946 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 14:58:15 crc kubenswrapper[4744]: I0106 14:58:15.954233 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-dfxz9"] Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.349122 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-m8c54" event={"ID":"4d50465b-09ea-4ff4-aa47-12f1c49ff207","Type":"ContainerStarted","Data":"10eb4cd86fc5b9c79d571ad70fd8f05c629458bd9f01ac45f97feefc587d4607"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.350945 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m88z9" event={"ID":"d9e793d1-65f3-43a8-914b-112690afc33f","Type":"ContainerStarted","Data":"70a155e38d038bc8d794f5c4229b8462818b4baae2a1afd8640b477bdf7710e3"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.350972 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m88z9" event={"ID":"d9e793d1-65f3-43a8-914b-112690afc33f","Type":"ContainerStarted","Data":"e1ae18628c2b1d69913b1b9e490361f02027a74b017bdbd6e8bdd7119016a621"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.352736 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vdnt5" event={"ID":"2b74efd3-9c21-4555-bfa8-5c1652b380dc","Type":"ContainerStarted","Data":"874175b57595591419a94ec9fceaadcc097a204bb4ce91d8ac4a450c636a87dc"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.355784 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" event={"ID":"987e985e-6a9f-4f21-8f8b-75ae7da8c45a","Type":"ContainerStarted","Data":"9049aec0e90a0cefb223a825ee5e1006c1cfd39580006f567b4f1539c501e00b"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.355833 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" event={"ID":"987e985e-6a9f-4f21-8f8b-75ae7da8c45a","Type":"ContainerStarted","Data":"f3db661f67316563e570e94ba820ec4f5c1880f94ecac654463b38567f1cc598"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.356967 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a","Type":"ContainerStarted","Data":"f1ecd45a43453fcf957dc964d93a7f7d299de053fc5b0a834c1733630a2f3e1f"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.358558 4744 generic.go:334] "Generic (PLEG): container finished" podID="52611203-15d5-41ba-b286-06d7058521d5" containerID="08d06bfbd539d1c9aa1ab2818d97faec0278eaf37aa82db0caa7b5e6ce49d915" exitCode=0 Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.358876 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" event={"ID":"52611203-15d5-41ba-b286-06d7058521d5","Type":"ContainerDied","Data":"08d06bfbd539d1c9aa1ab2818d97faec0278eaf37aa82db0caa7b5e6ce49d915"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.383064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dkghr" event={"ID":"d6889ed3-f69b-428a-a8f7-68e5ff3550df","Type":"ContainerStarted","Data":"2f399288422bab7afd8aec2f805890b5db03b5fcf6c7ac7b4d96ca779c4de80a"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.383118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dkghr" event={"ID":"d6889ed3-f69b-428a-a8f7-68e5ff3550df","Type":"ContainerStarted","Data":"e337059d831baf3b4eea4180f533eb4aad4c82d8e947ce30fbaa3a473a0b013e"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.389342 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dfxz9" event={"ID":"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b","Type":"ContainerStarted","Data":"b4ad122d47296d6847356b963f509d9d2d94a8744d46ea22e8e0d99d20fb22d0"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.400894 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-m88z9" podStartSLOduration=3.400872337 podStartE2EDuration="3.400872337s" podCreationTimestamp="2026-01-06 14:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:58:16.387447342 +0000 UTC m=+1293.014913670" watchObservedRunningTime="2026-01-06 14:58:16.400872337 +0000 UTC m=+1293.028338655" Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.419427 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbg7p" event={"ID":"e584b5b6-874e-48ea-b531-267a2c20562b","Type":"ContainerStarted","Data":"ef9e31b3df62352d36cf9b63fbc7eb73eaba47237a41cce64745243d49776a24"} Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.448632 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dkghr" podStartSLOduration=3.4486118 podStartE2EDuration="3.4486118s" podCreationTimestamp="2026-01-06 14:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:58:16.431354463 +0000 UTC m=+1293.058820791" watchObservedRunningTime="2026-01-06 14:58:16.4486118 +0000 UTC m=+1293.076078118" Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.868487 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.946984 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbzhw\" (UniqueName: \"kubernetes.io/projected/52611203-15d5-41ba-b286-06d7058521d5-kube-api-access-rbzhw\") pod \"52611203-15d5-41ba-b286-06d7058521d5\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.947097 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-swift-storage-0\") pod \"52611203-15d5-41ba-b286-06d7058521d5\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.947140 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-sb\") pod \"52611203-15d5-41ba-b286-06d7058521d5\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.947224 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-config\") pod \"52611203-15d5-41ba-b286-06d7058521d5\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.947404 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-svc\") pod \"52611203-15d5-41ba-b286-06d7058521d5\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.947476 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-nb\") pod \"52611203-15d5-41ba-b286-06d7058521d5\" (UID: \"52611203-15d5-41ba-b286-06d7058521d5\") " Jan 06 14:58:16 crc kubenswrapper[4744]: I0106 14:58:16.959565 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52611203-15d5-41ba-b286-06d7058521d5-kube-api-access-rbzhw" (OuterVolumeSpecName: "kube-api-access-rbzhw") pod "52611203-15d5-41ba-b286-06d7058521d5" (UID: "52611203-15d5-41ba-b286-06d7058521d5"). InnerVolumeSpecName "kube-api-access-rbzhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.030857 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "52611203-15d5-41ba-b286-06d7058521d5" (UID: "52611203-15d5-41ba-b286-06d7058521d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.043223 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "52611203-15d5-41ba-b286-06d7058521d5" (UID: "52611203-15d5-41ba-b286-06d7058521d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.052639 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "52611203-15d5-41ba-b286-06d7058521d5" (UID: "52611203-15d5-41ba-b286-06d7058521d5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.059962 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbzhw\" (UniqueName: \"kubernetes.io/projected/52611203-15d5-41ba-b286-06d7058521d5-kube-api-access-rbzhw\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.060013 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.060028 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.060040 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.112506 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "52611203-15d5-41ba-b286-06d7058521d5" (UID: "52611203-15d5-41ba-b286-06d7058521d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.134022 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-config" (OuterVolumeSpecName: "config") pod "52611203-15d5-41ba-b286-06d7058521d5" (UID: "52611203-15d5-41ba-b286-06d7058521d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.169841 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.169887 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52611203-15d5-41ba-b286-06d7058521d5-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.431486 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" event={"ID":"52611203-15d5-41ba-b286-06d7058521d5","Type":"ContainerDied","Data":"d044b772dc9ef968f2198e1177ce108d55f127731df181a7ce9e01738e6afb95"} Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.431552 4744 scope.go:117] "RemoveContainer" containerID="08d06bfbd539d1c9aa1ab2818d97faec0278eaf37aa82db0caa7b5e6ce49d915" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.431616 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-7xxr2" Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.434491 4744 generic.go:334] "Generic (PLEG): container finished" podID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" containerID="9049aec0e90a0cefb223a825ee5e1006c1cfd39580006f567b4f1539c501e00b" exitCode=0 Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.434537 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" event={"ID":"987e985e-6a9f-4f21-8f8b-75ae7da8c45a","Type":"ContainerDied","Data":"9049aec0e90a0cefb223a825ee5e1006c1cfd39580006f567b4f1539c501e00b"} Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.441400 4744 generic.go:334] "Generic (PLEG): container finished" podID="e584b5b6-874e-48ea-b531-267a2c20562b" containerID="ef9e31b3df62352d36cf9b63fbc7eb73eaba47237a41cce64745243d49776a24" exitCode=0 Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.442802 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbg7p" event={"ID":"e584b5b6-874e-48ea-b531-267a2c20562b","Type":"ContainerDied","Data":"ef9e31b3df62352d36cf9b63fbc7eb73eaba47237a41cce64745243d49776a24"} Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.520116 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.555240 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-7xxr2"] Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.566780 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-7xxr2"] Jan 06 14:58:17 crc kubenswrapper[4744]: I0106 14:58:17.724800 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52611203-15d5-41ba-b286-06d7058521d5" path="/var/lib/kubelet/pods/52611203-15d5-41ba-b286-06d7058521d5/volumes" Jan 06 14:58:19 crc kubenswrapper[4744]: I0106 14:58:19.485472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" event={"ID":"987e985e-6a9f-4f21-8f8b-75ae7da8c45a","Type":"ContainerStarted","Data":"7da164beb9c434748ceb41a8b419af03d092d18f8ec38a033f3931712201d946"} Jan 06 14:58:19 crc kubenswrapper[4744]: I0106 14:58:19.485978 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:19 crc kubenswrapper[4744]: I0106 14:58:19.522509 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" podStartSLOduration=6.522487237 podStartE2EDuration="6.522487237s" podCreationTimestamp="2026-01-06 14:58:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:58:19.514776463 +0000 UTC m=+1296.142242781" watchObservedRunningTime="2026-01-06 14:58:19.522487237 +0000 UTC m=+1296.149953555" Jan 06 14:58:24 crc kubenswrapper[4744]: I0106 14:58:24.692328 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:58:24 crc kubenswrapper[4744]: I0106 14:58:24.764750 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-f6xxt"] Jan 06 14:58:24 crc kubenswrapper[4744]: I0106 14:58:24.765015 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" podUID="def77701-47a7-4a7d-befc-75f5352554b6" containerName="dnsmasq-dns" containerID="cri-o://aa6faeaefb8d50626c37fa1461f08d96b9b5da11700a654c712adad5b879a3e0" gracePeriod=10 Jan 06 14:58:26 crc kubenswrapper[4744]: I0106 14:58:26.630402 4744 generic.go:334] "Generic (PLEG): container finished" podID="def77701-47a7-4a7d-befc-75f5352554b6" containerID="aa6faeaefb8d50626c37fa1461f08d96b9b5da11700a654c712adad5b879a3e0" exitCode=0 Jan 06 14:58:26 crc kubenswrapper[4744]: I0106 14:58:26.630463 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" event={"ID":"def77701-47a7-4a7d-befc-75f5352554b6","Type":"ContainerDied","Data":"aa6faeaefb8d50626c37fa1461f08d96b9b5da11700a654c712adad5b879a3e0"} Jan 06 14:58:27 crc kubenswrapper[4744]: I0106 14:58:27.684242 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" podUID="def77701-47a7-4a7d-befc-75f5352554b6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: connect: connection refused" Jan 06 14:58:29 crc kubenswrapper[4744]: E0106 14:58:29.752763 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Jan 06 14:58:29 crc kubenswrapper[4744]: E0106 14:58:29.753711 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-62p9x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-6ltjv_openstack(5ae28705-95b4-4dd8-ab04-bb1e7202ae6b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:58:29 crc kubenswrapper[4744]: E0106 14:58:29.754871 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-6ltjv" podUID="5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" Jan 06 14:58:30 crc kubenswrapper[4744]: E0106 14:58:30.675663 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-6ltjv" podUID="5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" Jan 06 14:58:32 crc kubenswrapper[4744]: I0106 14:58:32.697359 4744 generic.go:334] "Generic (PLEG): container finished" podID="d6889ed3-f69b-428a-a8f7-68e5ff3550df" containerID="2f399288422bab7afd8aec2f805890b5db03b5fcf6c7ac7b4d96ca779c4de80a" exitCode=0 Jan 06 14:58:32 crc kubenswrapper[4744]: I0106 14:58:32.697443 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dkghr" event={"ID":"d6889ed3-f69b-428a-a8f7-68e5ff3550df","Type":"ContainerDied","Data":"2f399288422bab7afd8aec2f805890b5db03b5fcf6c7ac7b4d96ca779c4de80a"} Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.329967 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.409070 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-swift-storage-0\") pod \"def77701-47a7-4a7d-befc-75f5352554b6\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.409253 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-nb\") pod \"def77701-47a7-4a7d-befc-75f5352554b6\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.409318 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n7kq\" (UniqueName: \"kubernetes.io/projected/def77701-47a7-4a7d-befc-75f5352554b6-kube-api-access-7n7kq\") pod \"def77701-47a7-4a7d-befc-75f5352554b6\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.409437 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-config\") pod \"def77701-47a7-4a7d-befc-75f5352554b6\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.410490 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-sb\") pod \"def77701-47a7-4a7d-befc-75f5352554b6\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.410617 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-svc\") pod \"def77701-47a7-4a7d-befc-75f5352554b6\" (UID: \"def77701-47a7-4a7d-befc-75f5352554b6\") " Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.437870 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/def77701-47a7-4a7d-befc-75f5352554b6-kube-api-access-7n7kq" (OuterVolumeSpecName: "kube-api-access-7n7kq") pod "def77701-47a7-4a7d-befc-75f5352554b6" (UID: "def77701-47a7-4a7d-befc-75f5352554b6"). InnerVolumeSpecName "kube-api-access-7n7kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.461082 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "def77701-47a7-4a7d-befc-75f5352554b6" (UID: "def77701-47a7-4a7d-befc-75f5352554b6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.465694 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "def77701-47a7-4a7d-befc-75f5352554b6" (UID: "def77701-47a7-4a7d-befc-75f5352554b6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.467504 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "def77701-47a7-4a7d-befc-75f5352554b6" (UID: "def77701-47a7-4a7d-befc-75f5352554b6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.475415 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "def77701-47a7-4a7d-befc-75f5352554b6" (UID: "def77701-47a7-4a7d-befc-75f5352554b6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.496945 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-config" (OuterVolumeSpecName: "config") pod "def77701-47a7-4a7d-befc-75f5352554b6" (UID: "def77701-47a7-4a7d-befc-75f5352554b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.513956 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.513988 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.514001 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.514011 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.514021 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/def77701-47a7-4a7d-befc-75f5352554b6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.514030 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n7kq\" (UniqueName: \"kubernetes.io/projected/def77701-47a7-4a7d-befc-75f5352554b6-kube-api-access-7n7kq\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.717482 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" event={"ID":"def77701-47a7-4a7d-befc-75f5352554b6","Type":"ContainerDied","Data":"92a7d5c0e51adeea5d167f482368e090b261027c8e56936f2f584815f311d820"} Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.717705 4744 scope.go:117] "RemoveContainer" containerID="aa6faeaefb8d50626c37fa1461f08d96b9b5da11700a654c712adad5b879a3e0" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.717792 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.771606 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-f6xxt"] Jan 06 14:58:34 crc kubenswrapper[4744]: I0106 14:58:34.784975 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-f6xxt"] Jan 06 14:58:35 crc kubenswrapper[4744]: I0106 14:58:35.724982 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="def77701-47a7-4a7d-befc-75f5352554b6" path="/var/lib/kubelet/pods/def77701-47a7-4a7d-befc-75f5352554b6/volumes" Jan 06 14:58:36 crc kubenswrapper[4744]: E0106 14:58:36.431051 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Jan 06 14:58:36 crc kubenswrapper[4744]: E0106 14:58:36.431513 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5n6zx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-dfxz9_openstack(be9d6eb2-8ad6-4eb4-9208-a89adf751c4b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:58:36 crc kubenswrapper[4744]: E0106 14:58:36.432754 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-dfxz9" podUID="be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" Jan 06 14:58:36 crc kubenswrapper[4744]: E0106 14:58:36.739894 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-dfxz9" podUID="be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" Jan 06 14:58:37 crc kubenswrapper[4744]: I0106 14:58:37.684874 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-f6xxt" podUID="def77701-47a7-4a7d-befc-75f5352554b6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: i/o timeout" Jan 06 14:58:37 crc kubenswrapper[4744]: I0106 14:58:37.758067 4744 generic.go:334] "Generic (PLEG): container finished" podID="cb6ad61a-8190-4d4d-987c-f609c1e8cf44" containerID="2ab20d9a010b2164f31b5440bef346b4bb35d8a31cc6ba96c6bfba16eaf5f6b4" exitCode=0 Jan 06 14:58:37 crc kubenswrapper[4744]: I0106 14:58:37.758111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-68rr4" event={"ID":"cb6ad61a-8190-4d4d-987c-f609c1e8cf44","Type":"ContainerDied","Data":"2ab20d9a010b2164f31b5440bef346b4bb35d8a31cc6ba96c6bfba16eaf5f6b4"} Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.637006 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.734638 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-config-data\") pod \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.734986 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-combined-ca-bundle\") pod \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.735028 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-credential-keys\") pod \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.735093 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-fernet-keys\") pod \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.735218 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w8xs\" (UniqueName: \"kubernetes.io/projected/d6889ed3-f69b-428a-a8f7-68e5ff3550df-kube-api-access-6w8xs\") pod \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.735291 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-scripts\") pod \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\" (UID: \"d6889ed3-f69b-428a-a8f7-68e5ff3550df\") " Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.751393 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d6889ed3-f69b-428a-a8f7-68e5ff3550df" (UID: "d6889ed3-f69b-428a-a8f7-68e5ff3550df"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.754808 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6889ed3-f69b-428a-a8f7-68e5ff3550df-kube-api-access-6w8xs" (OuterVolumeSpecName: "kube-api-access-6w8xs") pod "d6889ed3-f69b-428a-a8f7-68e5ff3550df" (UID: "d6889ed3-f69b-428a-a8f7-68e5ff3550df"). InnerVolumeSpecName "kube-api-access-6w8xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.755833 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d6889ed3-f69b-428a-a8f7-68e5ff3550df" (UID: "d6889ed3-f69b-428a-a8f7-68e5ff3550df"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.756116 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-scripts" (OuterVolumeSpecName: "scripts") pod "d6889ed3-f69b-428a-a8f7-68e5ff3550df" (UID: "d6889ed3-f69b-428a-a8f7-68e5ff3550df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.766420 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6889ed3-f69b-428a-a8f7-68e5ff3550df" (UID: "d6889ed3-f69b-428a-a8f7-68e5ff3550df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.769393 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-config-data" (OuterVolumeSpecName: "config-data") pod "d6889ed3-f69b-428a-a8f7-68e5ff3550df" (UID: "d6889ed3-f69b-428a-a8f7-68e5ff3550df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.838195 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.838234 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.838245 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.838254 4744 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.838264 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d6889ed3-f69b-428a-a8f7-68e5ff3550df-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.838284 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w8xs\" (UniqueName: \"kubernetes.io/projected/d6889ed3-f69b-428a-a8f7-68e5ff3550df-kube-api-access-6w8xs\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.907992 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dkghr" event={"ID":"d6889ed3-f69b-428a-a8f7-68e5ff3550df","Type":"ContainerDied","Data":"e337059d831baf3b4eea4180f533eb4aad4c82d8e947ce30fbaa3a473a0b013e"} Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.908041 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e337059d831baf3b4eea4180f533eb4aad4c82d8e947ce30fbaa3a473a0b013e" Jan 06 14:58:50 crc kubenswrapper[4744]: I0106 14:58:50.908897 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dkghr" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.131240 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.131535 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad,Command:[/bin/opm],Args:[serve /extracted-catalog/catalog --cache-dir=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOMEMLIMIT,Value:30MiB,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{31457280 0} {} 30Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hmfcw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-pbg7p_openshift-marketplace(e584b5b6-874e-48ea-b531-267a2c20562b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.132815 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-pbg7p" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.173897 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.174124 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7gbb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-m8c54_openstack(4d50465b-09ea-4ff4-aa47-12f1c49ff207): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.175352 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-m8c54" podUID="4d50465b-09ea-4ff4-aa47-12f1c49ff207" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.619375 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-68rr4" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.619861 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.619991 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd5h5f4hd7h4h579h696h567h649h9hb8h5c6h588h5d6h656h656h656h5dh66h64dh56chc9hdfh685h8dh68h669h54h55hcdh57ch5hbq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ktcb5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(0832cc9a-5d34-4e11-8d3b-cc7c0674b73a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.662042 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-db-sync-config-data\") pod \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.662111 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hstt2\" (UniqueName: \"kubernetes.io/projected/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-kube-api-access-hstt2\") pod \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.662294 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-combined-ca-bundle\") pod \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.662381 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-config-data\") pod \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.680543 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-kube-api-access-hstt2" (OuterVolumeSpecName: "kube-api-access-hstt2") pod "cb6ad61a-8190-4d4d-987c-f609c1e8cf44" (UID: "cb6ad61a-8190-4d4d-987c-f609c1e8cf44"). InnerVolumeSpecName "kube-api-access-hstt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.682044 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cb6ad61a-8190-4d4d-987c-f609c1e8cf44" (UID: "cb6ad61a-8190-4d4d-987c-f609c1e8cf44"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.742382 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cb6ad61a-8190-4d4d-987c-f609c1e8cf44" (UID: "cb6ad61a-8190-4d4d-987c-f609c1e8cf44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.763733 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-config-data" (OuterVolumeSpecName: "config-data") pod "cb6ad61a-8190-4d4d-987c-f609c1e8cf44" (UID: "cb6ad61a-8190-4d4d-987c-f609c1e8cf44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.764193 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-config-data\") pod \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\" (UID: \"cb6ad61a-8190-4d4d-987c-f609c1e8cf44\") " Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.765070 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.765094 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.765107 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hstt2\" (UniqueName: \"kubernetes.io/projected/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-kube-api-access-hstt2\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:51 crc kubenswrapper[4744]: W0106 14:58:51.765500 4744 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/cb6ad61a-8190-4d4d-987c-f609c1e8cf44/volumes/kubernetes.io~secret/config-data Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.765520 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-config-data" (OuterVolumeSpecName: "config-data") pod "cb6ad61a-8190-4d4d-987c-f609c1e8cf44" (UID: "cb6ad61a-8190-4d4d-987c-f609c1e8cf44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.836634 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dkghr"] Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.836905 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dkghr"] Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.844363 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-tffgx"] Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.845321 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52611203-15d5-41ba-b286-06d7058521d5" containerName="init" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.845371 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="52611203-15d5-41ba-b286-06d7058521d5" containerName="init" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.845397 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6889ed3-f69b-428a-a8f7-68e5ff3550df" containerName="keystone-bootstrap" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.845459 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6889ed3-f69b-428a-a8f7-68e5ff3550df" containerName="keystone-bootstrap" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.845484 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def77701-47a7-4a7d-befc-75f5352554b6" containerName="init" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.845494 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="def77701-47a7-4a7d-befc-75f5352554b6" containerName="init" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.845543 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def77701-47a7-4a7d-befc-75f5352554b6" containerName="dnsmasq-dns" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.845553 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="def77701-47a7-4a7d-befc-75f5352554b6" containerName="dnsmasq-dns" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.845568 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb6ad61a-8190-4d4d-987c-f609c1e8cf44" containerName="glance-db-sync" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.845576 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb6ad61a-8190-4d4d-987c-f609c1e8cf44" containerName="glance-db-sync" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.845981 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb6ad61a-8190-4d4d-987c-f609c1e8cf44" containerName="glance-db-sync" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.846724 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="def77701-47a7-4a7d-befc-75f5352554b6" containerName="dnsmasq-dns" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.846785 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6889ed3-f69b-428a-a8f7-68e5ff3550df" containerName="keystone-bootstrap" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.846800 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="52611203-15d5-41ba-b286-06d7058521d5" containerName="init" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.847872 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.850801 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.851839 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.852043 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.852071 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.852081 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8hqr4" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.858619 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tffgx"] Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.869796 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cb6ad61a-8190-4d4d-987c-f609c1e8cf44-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.920369 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-68rr4" event={"ID":"cb6ad61a-8190-4d4d-987c-f609c1e8cf44","Type":"ContainerDied","Data":"3dd926f3cf954db0de06835d3632a80433b3a1b42f8f4c0ca87fcf7508d93ef3"} Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.920443 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dd926f3cf954db0de06835d3632a80433b3a1b42f8f4c0ca87fcf7508d93ef3" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.920500 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-68rr4" Jan 06 14:58:51 crc kubenswrapper[4744]: E0106 14:58:51.923707 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-m8c54" podUID="4d50465b-09ea-4ff4-aa47-12f1c49ff207" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.971978 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-config-data\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.972023 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-fernet-keys\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.972047 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-credential-keys\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.972222 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-scripts\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.972272 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-combined-ca-bundle\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:51 crc kubenswrapper[4744]: I0106 14:58:51.972319 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp7pk\" (UniqueName: \"kubernetes.io/projected/8c5385b7-40c0-4927-87fe-3db7adf6cdab-kube-api-access-hp7pk\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.074127 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-scripts\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.074215 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-combined-ca-bundle\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.074275 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp7pk\" (UniqueName: \"kubernetes.io/projected/8c5385b7-40c0-4927-87fe-3db7adf6cdab-kube-api-access-hp7pk\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.074293 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-config-data\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.074315 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-fernet-keys\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.074335 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-credential-keys\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.078313 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-scripts\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.078793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-fernet-keys\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.078793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-credential-keys\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.080031 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-config-data\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.082655 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-combined-ca-bundle\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.096701 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp7pk\" (UniqueName: \"kubernetes.io/projected/8c5385b7-40c0-4927-87fe-3db7adf6cdab-kube-api-access-hp7pk\") pod \"keystone-bootstrap-tffgx\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:52 crc kubenswrapper[4744]: I0106 14:58:52.167103 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.100801 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-gndpz"] Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.109345 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.143371 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-gndpz"] Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.198837 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.198929 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.199019 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28xgr\" (UniqueName: \"kubernetes.io/projected/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-kube-api-access-28xgr\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.199043 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.199174 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-config\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.199212 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.300564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-config\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.300611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.300685 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.300728 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.300789 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28xgr\" (UniqueName: \"kubernetes.io/projected/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-kube-api-access-28xgr\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.300806 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.301725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.301750 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-config\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.301906 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.302815 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.303862 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.335302 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28xgr\" (UniqueName: \"kubernetes.io/projected/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-kube-api-access-28xgr\") pod \"dnsmasq-dns-57c957c4ff-gndpz\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.463057 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.725089 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6889ed3-f69b-428a-a8f7-68e5ff3550df" path="/var/lib/kubelet/pods/d6889ed3-f69b-428a-a8f7-68e5ff3550df/volumes" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.966228 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.969123 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.971570 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.971609 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lxm4v" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.972753 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 06 14:58:53 crc kubenswrapper[4744]: I0106 14:58:53.976929 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.016943 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.016992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.017025 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28w4l\" (UniqueName: \"kubernetes.io/projected/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-kube-api-access-28w4l\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.017062 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-config-data\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.017083 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.017098 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-logs\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.017125 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-scripts\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.118984 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.119061 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28w4l\" (UniqueName: \"kubernetes.io/projected/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-kube-api-access-28w4l\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.119114 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-config-data\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.119141 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.119184 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-logs\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.119226 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-scripts\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.119395 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.119908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.120392 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-logs\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.122573 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.122611 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ac0041d382eb549e2b0d342e4bfe43b8e025e345c53521653e27358ce8689034/globalmount\"" pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.130805 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-scripts\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.130918 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.147986 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-config-data\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.186362 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28w4l\" (UniqueName: \"kubernetes.io/projected/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-kube-api-access-28w4l\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.244673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.260031 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.262951 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.266765 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.288665 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.297258 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.329082 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.329172 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-logs\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.329203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.329228 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.329427 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28rb6\" (UniqueName: \"kubernetes.io/projected/f40b2a66-a5cd-4952-ac90-947bb0b46c28-kube-api-access-28rb6\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.329480 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.329907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.431970 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.432151 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.432234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.432269 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-logs\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.432298 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.432320 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.432412 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28rb6\" (UniqueName: \"kubernetes.io/projected/f40b2a66-a5cd-4952-ac90-947bb0b46c28-kube-api-access-28rb6\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.433075 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.433220 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-logs\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.437997 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.438203 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a751ba72a1be54984524758ff1fa5b9a696d9d8d73f1feb0e198d80940b6b3a7/globalmount\"" pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.439397 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.441864 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.445994 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.452257 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28rb6\" (UniqueName: \"kubernetes.io/projected/f40b2a66-a5cd-4952-ac90-947bb0b46c28-kube-api-access-28rb6\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.488177 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:58:54 crc kubenswrapper[4744]: I0106 14:58:54.593591 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 14:58:55 crc kubenswrapper[4744]: E0106 14:58:55.039103 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"\"" pod="openshift-marketplace/redhat-operators-pbg7p" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" Jan 06 14:58:55 crc kubenswrapper[4744]: I0106 14:58:55.050385 4744 scope.go:117] "RemoveContainer" containerID="87e558bdac3b358a19862e5588c16cf2df9999b4b7ff455ac1e4e3d649928669" Jan 06 14:58:55 crc kubenswrapper[4744]: E0106 14:58:55.085754 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Jan 06 14:58:55 crc kubenswrapper[4744]: E0106 14:58:55.085998 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bbzxc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-vdnt5_openstack(2b74efd3-9c21-4555-bfa8-5c1652b380dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 14:58:55 crc kubenswrapper[4744]: E0106 14:58:55.087185 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-vdnt5" podUID="2b74efd3-9c21-4555-bfa8-5c1652b380dc" Jan 06 14:58:55 crc kubenswrapper[4744]: I0106 14:58:55.540272 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 14:58:55 crc kubenswrapper[4744]: I0106 14:58:55.593437 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 14:58:56 crc kubenswrapper[4744]: E0106 14:58:56.001881 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-vdnt5" podUID="2b74efd3-9c21-4555-bfa8-5c1652b380dc" Jan 06 14:58:56 crc kubenswrapper[4744]: I0106 14:58:56.392941 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-tffgx"] Jan 06 14:58:56 crc kubenswrapper[4744]: I0106 14:58:56.503009 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 14:58:56 crc kubenswrapper[4744]: I0106 14:58:56.596725 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-gndpz"] Jan 06 14:58:56 crc kubenswrapper[4744]: I0106 14:58:56.798767 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 14:58:56 crc kubenswrapper[4744]: W0106 14:58:56.967853 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd69c6e25_d85a_4ac9_ab43_17ee2752e5be.slice/crio-4fd51258a1d3082bcce99cd582769e15eab469a37abb443fdf8f56b8f59cd152 WatchSource:0}: Error finding container 4fd51258a1d3082bcce99cd582769e15eab469a37abb443fdf8f56b8f59cd152: Status 404 returned error can't find the container with id 4fd51258a1d3082bcce99cd582769e15eab469a37abb443fdf8f56b8f59cd152 Jan 06 14:58:57 crc kubenswrapper[4744]: I0106 14:58:57.027548 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tffgx" event={"ID":"8c5385b7-40c0-4927-87fe-3db7adf6cdab","Type":"ContainerStarted","Data":"daf852d856ecda8349f1971d99249fc85f011bcea2ceec16a4e85527881a8259"} Jan 06 14:58:57 crc kubenswrapper[4744]: I0106 14:58:57.031992 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6ltjv" event={"ID":"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b","Type":"ContainerStarted","Data":"614fe98476c53fb750be24913e7fb2bb300d66d103e2b2ab18c9ac222ce7b9e2"} Jan 06 14:58:57 crc kubenswrapper[4744]: I0106 14:58:57.034082 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d69c6e25-d85a-4ac9-ab43-17ee2752e5be","Type":"ContainerStarted","Data":"4fd51258a1d3082bcce99cd582769e15eab469a37abb443fdf8f56b8f59cd152"} Jan 06 14:58:57 crc kubenswrapper[4744]: I0106 14:58:57.045594 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dfxz9" event={"ID":"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b","Type":"ContainerStarted","Data":"6dbf2f5d4c23d880674670d0d4a0482afb770c9fd6ec883381a24e1ed079b285"} Jan 06 14:58:57 crc kubenswrapper[4744]: I0106 14:58:57.052823 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f40b2a66-a5cd-4952-ac90-947bb0b46c28","Type":"ContainerStarted","Data":"faf739dc7477888e7c38e12e6adc88fb5a197a45f77b140bcd45105c774dbc55"} Jan 06 14:58:57 crc kubenswrapper[4744]: I0106 14:58:57.054819 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" event={"ID":"e39dd03e-1fc4-4650-b0b6-d0067a81ed77","Type":"ContainerStarted","Data":"9a0de90f94af9faad3fed0d2239bbf58f5c9b3dc9346c05b9fe896d3f8611b3b"} Jan 06 14:58:57 crc kubenswrapper[4744]: I0106 14:58:57.071469 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-6ltjv" podStartSLOduration=3.446894944 podStartE2EDuration="44.071453505s" podCreationTimestamp="2026-01-06 14:58:13 +0000 UTC" firstStartedPulling="2026-01-06 14:58:15.147498392 +0000 UTC m=+1291.774964710" lastFinishedPulling="2026-01-06 14:58:55.772056953 +0000 UTC m=+1332.399523271" observedRunningTime="2026-01-06 14:58:57.058341258 +0000 UTC m=+1333.685807576" watchObservedRunningTime="2026-01-06 14:58:57.071453505 +0000 UTC m=+1333.698919823" Jan 06 14:58:57 crc kubenswrapper[4744]: I0106 14:58:57.086276 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-dfxz9" podStartSLOduration=3.256403041 podStartE2EDuration="43.086256797s" podCreationTimestamp="2026-01-06 14:58:14 +0000 UTC" firstStartedPulling="2026-01-06 14:58:15.942139276 +0000 UTC m=+1292.569605594" lastFinishedPulling="2026-01-06 14:58:55.771993032 +0000 UTC m=+1332.399459350" observedRunningTime="2026-01-06 14:58:57.079239961 +0000 UTC m=+1333.706706279" watchObservedRunningTime="2026-01-06 14:58:57.086256797 +0000 UTC m=+1333.713723115" Jan 06 14:58:58 crc kubenswrapper[4744]: I0106 14:58:58.076069 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a","Type":"ContainerStarted","Data":"61047799c5a70e9d54f5a34f6084719581dd169ba38e6826d47527d3b52e0f41"} Jan 06 14:58:58 crc kubenswrapper[4744]: I0106 14:58:58.078359 4744 generic.go:334] "Generic (PLEG): container finished" podID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" containerID="0f83d8110bf3bd2c85ca8494886b4c9d3b38d50825e4441c41bdc97998490136" exitCode=0 Jan 06 14:58:58 crc kubenswrapper[4744]: I0106 14:58:58.078413 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" event={"ID":"e39dd03e-1fc4-4650-b0b6-d0067a81ed77","Type":"ContainerDied","Data":"0f83d8110bf3bd2c85ca8494886b4c9d3b38d50825e4441c41bdc97998490136"} Jan 06 14:58:58 crc kubenswrapper[4744]: I0106 14:58:58.092495 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tffgx" event={"ID":"8c5385b7-40c0-4927-87fe-3db7adf6cdab","Type":"ContainerStarted","Data":"c948c1d0335191aa70e62006bf06381bf3b03ffdaa262d1ef0601687003ff060"} Jan 06 14:58:58 crc kubenswrapper[4744]: I0106 14:58:58.095277 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f40b2a66-a5cd-4952-ac90-947bb0b46c28","Type":"ContainerStarted","Data":"44fbad7848baba929d60dee2bbd48d3085b46e976ece76ea3117345a243dc919"} Jan 06 14:58:58 crc kubenswrapper[4744]: I0106 14:58:58.133061 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-tffgx" podStartSLOduration=7.133044969 podStartE2EDuration="7.133044969s" podCreationTimestamp="2026-01-06 14:58:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:58:58.123657821 +0000 UTC m=+1334.751124139" watchObservedRunningTime="2026-01-06 14:58:58.133044969 +0000 UTC m=+1334.760511287" Jan 06 14:58:59 crc kubenswrapper[4744]: I0106 14:58:59.110203 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d69c6e25-d85a-4ac9-ab43-17ee2752e5be","Type":"ContainerStarted","Data":"032bad6126a09cc22dac483eebb2a7927f98b259f49ef8b33006a6e9d73613f5"} Jan 06 14:58:59 crc kubenswrapper[4744]: I0106 14:58:59.113479 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f40b2a66-a5cd-4952-ac90-947bb0b46c28","Type":"ContainerStarted","Data":"48cbf5a5016c23ffdaf2ff5a591e86f1648709cf93b8f2dc9a239fb778663832"} Jan 06 14:58:59 crc kubenswrapper[4744]: I0106 14:58:59.113654 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerName="glance-log" containerID="cri-o://44fbad7848baba929d60dee2bbd48d3085b46e976ece76ea3117345a243dc919" gracePeriod=30 Jan 06 14:58:59 crc kubenswrapper[4744]: I0106 14:58:59.113653 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerName="glance-httpd" containerID="cri-o://48cbf5a5016c23ffdaf2ff5a591e86f1648709cf93b8f2dc9a239fb778663832" gracePeriod=30 Jan 06 14:58:59 crc kubenswrapper[4744]: I0106 14:58:59.140656 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.140636515 podStartE2EDuration="6.140636515s" podCreationTimestamp="2026-01-06 14:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:58:59.138388276 +0000 UTC m=+1335.765854654" watchObservedRunningTime="2026-01-06 14:58:59.140636515 +0000 UTC m=+1335.768102833" Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.125032 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d69c6e25-d85a-4ac9-ab43-17ee2752e5be","Type":"ContainerStarted","Data":"13782914df07095b238be4ab6d8eb719b7d53c1314b598f8b307e3c0b5879b3b"} Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.125533 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerName="glance-log" containerID="cri-o://032bad6126a09cc22dac483eebb2a7927f98b259f49ef8b33006a6e9d73613f5" gracePeriod=30 Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.126221 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerName="glance-httpd" containerID="cri-o://13782914df07095b238be4ab6d8eb719b7d53c1314b598f8b307e3c0b5879b3b" gracePeriod=30 Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.128575 4744 generic.go:334] "Generic (PLEG): container finished" podID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerID="48cbf5a5016c23ffdaf2ff5a591e86f1648709cf93b8f2dc9a239fb778663832" exitCode=143 Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.128601 4744 generic.go:334] "Generic (PLEG): container finished" podID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerID="44fbad7848baba929d60dee2bbd48d3085b46e976ece76ea3117345a243dc919" exitCode=143 Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.128643 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f40b2a66-a5cd-4952-ac90-947bb0b46c28","Type":"ContainerDied","Data":"48cbf5a5016c23ffdaf2ff5a591e86f1648709cf93b8f2dc9a239fb778663832"} Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.128671 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f40b2a66-a5cd-4952-ac90-947bb0b46c28","Type":"ContainerDied","Data":"44fbad7848baba929d60dee2bbd48d3085b46e976ece76ea3117345a243dc919"} Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.136541 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" event={"ID":"e39dd03e-1fc4-4650-b0b6-d0067a81ed77","Type":"ContainerStarted","Data":"c8f1d227f184e3c8f79983e9b7dd37aafb20ed8b3327dc627f8208edcd1b3001"} Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.136682 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.163838 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.163817383 podStartE2EDuration="8.163817383s" podCreationTimestamp="2026-01-06 14:58:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:59:00.147114241 +0000 UTC m=+1336.774580559" watchObservedRunningTime="2026-01-06 14:59:00.163817383 +0000 UTC m=+1336.791283701" Jan 06 14:59:00 crc kubenswrapper[4744]: I0106 14:59:00.184026 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" podStartSLOduration=7.184005336 podStartE2EDuration="7.184005336s" podCreationTimestamp="2026-01-06 14:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:59:00.172368649 +0000 UTC m=+1336.799834967" watchObservedRunningTime="2026-01-06 14:59:00.184005336 +0000 UTC m=+1336.811471654" Jan 06 14:59:01 crc kubenswrapper[4744]: I0106 14:59:01.152061 4744 generic.go:334] "Generic (PLEG): container finished" podID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerID="13782914df07095b238be4ab6d8eb719b7d53c1314b598f8b307e3c0b5879b3b" exitCode=0 Jan 06 14:59:01 crc kubenswrapper[4744]: I0106 14:59:01.152427 4744 generic.go:334] "Generic (PLEG): container finished" podID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerID="032bad6126a09cc22dac483eebb2a7927f98b259f49ef8b33006a6e9d73613f5" exitCode=143 Jan 06 14:59:01 crc kubenswrapper[4744]: I0106 14:59:01.153603 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d69c6e25-d85a-4ac9-ab43-17ee2752e5be","Type":"ContainerDied","Data":"13782914df07095b238be4ab6d8eb719b7d53c1314b598f8b307e3c0b5879b3b"} Jan 06 14:59:01 crc kubenswrapper[4744]: I0106 14:59:01.153652 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d69c6e25-d85a-4ac9-ab43-17ee2752e5be","Type":"ContainerDied","Data":"032bad6126a09cc22dac483eebb2a7927f98b259f49ef8b33006a6e9d73613f5"} Jan 06 14:59:08 crc kubenswrapper[4744]: I0106 14:59:08.465408 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 14:59:08 crc kubenswrapper[4744]: I0106 14:59:08.570448 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-xm2rp"] Jan 06 14:59:08 crc kubenswrapper[4744]: I0106 14:59:08.570746 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" podUID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" containerName="dnsmasq-dns" containerID="cri-o://7da164beb9c434748ceb41a8b419af03d092d18f8ec38a033f3931712201d946" gracePeriod=10 Jan 06 14:59:09 crc kubenswrapper[4744]: I0106 14:59:09.320338 4744 generic.go:334] "Generic (PLEG): container finished" podID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" containerID="7da164beb9c434748ceb41a8b419af03d092d18f8ec38a033f3931712201d946" exitCode=0 Jan 06 14:59:09 crc kubenswrapper[4744]: I0106 14:59:09.320418 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" event={"ID":"987e985e-6a9f-4f21-8f8b-75ae7da8c45a","Type":"ContainerDied","Data":"7da164beb9c434748ceb41a8b419af03d092d18f8ec38a033f3931712201d946"} Jan 06 14:59:09 crc kubenswrapper[4744]: I0106 14:59:09.691421 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" podUID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.183:5353: connect: connection refused" Jan 06 14:59:10 crc kubenswrapper[4744]: I0106 14:59:10.337877 4744 generic.go:334] "Generic (PLEG): container finished" podID="8c5385b7-40c0-4927-87fe-3db7adf6cdab" containerID="c948c1d0335191aa70e62006bf06381bf3b03ffdaa262d1ef0601687003ff060" exitCode=0 Jan 06 14:59:10 crc kubenswrapper[4744]: I0106 14:59:10.337950 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tffgx" event={"ID":"8c5385b7-40c0-4927-87fe-3db7adf6cdab","Type":"ContainerDied","Data":"c948c1d0335191aa70e62006bf06381bf3b03ffdaa262d1ef0601687003ff060"} Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.043101 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.069102 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.073859 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201095 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-combined-ca-bundle\") pod \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201225 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28rb6\" (UniqueName: \"kubernetes.io/projected/f40b2a66-a5cd-4952-ac90-947bb0b46c28-kube-api-access-28rb6\") pod \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201283 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-config-data\") pod \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201301 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28w4l\" (UniqueName: \"kubernetes.io/projected/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-kube-api-access-28w4l\") pod \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201341 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-fernet-keys\") pod \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201380 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-config-data\") pod \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201437 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-logs\") pod \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201468 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-scripts\") pod \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201493 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-logs\") pod \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201535 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-combined-ca-bundle\") pod \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201704 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201742 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-combined-ca-bundle\") pod \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201768 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-config-data\") pod \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201796 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-httpd-run\") pod \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201823 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-scripts\") pod \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201909 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201942 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp7pk\" (UniqueName: \"kubernetes.io/projected/8c5385b7-40c0-4927-87fe-3db7adf6cdab-kube-api-access-hp7pk\") pod \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201969 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-scripts\") pod \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\" (UID: \"d69c6e25-d85a-4ac9-ab43-17ee2752e5be\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201988 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-logs" (OuterVolumeSpecName: "logs") pod "f40b2a66-a5cd-4952-ac90-947bb0b46c28" (UID: "f40b2a66-a5cd-4952-ac90-947bb0b46c28"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201984 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-logs" (OuterVolumeSpecName: "logs") pod "d69c6e25-d85a-4ac9-ab43-17ee2752e5be" (UID: "d69c6e25-d85a-4ac9-ab43-17ee2752e5be"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.201997 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-httpd-run\") pod \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\" (UID: \"f40b2a66-a5cd-4952-ac90-947bb0b46c28\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.202069 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-credential-keys\") pod \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\" (UID: \"8c5385b7-40c0-4927-87fe-3db7adf6cdab\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.202194 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f40b2a66-a5cd-4952-ac90-947bb0b46c28" (UID: "f40b2a66-a5cd-4952-ac90-947bb0b46c28"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.202246 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d69c6e25-d85a-4ac9-ab43-17ee2752e5be" (UID: "d69c6e25-d85a-4ac9-ab43-17ee2752e5be"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.202623 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-logs\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.202646 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-logs\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.202656 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.202665 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f40b2a66-a5cd-4952-ac90-947bb0b46c28-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.208414 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f40b2a66-a5cd-4952-ac90-947bb0b46c28-kube-api-access-28rb6" (OuterVolumeSpecName: "kube-api-access-28rb6") pod "f40b2a66-a5cd-4952-ac90-947bb0b46c28" (UID: "f40b2a66-a5cd-4952-ac90-947bb0b46c28"). InnerVolumeSpecName "kube-api-access-28rb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.208979 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-scripts" (OuterVolumeSpecName: "scripts") pod "f40b2a66-a5cd-4952-ac90-947bb0b46c28" (UID: "f40b2a66-a5cd-4952-ac90-947bb0b46c28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.211259 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8c5385b7-40c0-4927-87fe-3db7adf6cdab" (UID: "8c5385b7-40c0-4927-87fe-3db7adf6cdab"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.211313 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-scripts" (OuterVolumeSpecName: "scripts") pod "8c5385b7-40c0-4927-87fe-3db7adf6cdab" (UID: "8c5385b7-40c0-4927-87fe-3db7adf6cdab"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.211342 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8c5385b7-40c0-4927-87fe-3db7adf6cdab" (UID: "8c5385b7-40c0-4927-87fe-3db7adf6cdab"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.213356 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-kube-api-access-28w4l" (OuterVolumeSpecName: "kube-api-access-28w4l") pod "d69c6e25-d85a-4ac9-ab43-17ee2752e5be" (UID: "d69c6e25-d85a-4ac9-ab43-17ee2752e5be"). InnerVolumeSpecName "kube-api-access-28w4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.215343 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c5385b7-40c0-4927-87fe-3db7adf6cdab-kube-api-access-hp7pk" (OuterVolumeSpecName: "kube-api-access-hp7pk") pod "8c5385b7-40c0-4927-87fe-3db7adf6cdab" (UID: "8c5385b7-40c0-4927-87fe-3db7adf6cdab"). InnerVolumeSpecName "kube-api-access-hp7pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.225968 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-scripts" (OuterVolumeSpecName: "scripts") pod "d69c6e25-d85a-4ac9-ab43-17ee2752e5be" (UID: "d69c6e25-d85a-4ac9-ab43-17ee2752e5be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.227397 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b" (OuterVolumeSpecName: "glance") pod "f40b2a66-a5cd-4952-ac90-947bb0b46c28" (UID: "f40b2a66-a5cd-4952-ac90-947bb0b46c28"). InnerVolumeSpecName "pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.243195 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c5385b7-40c0-4927-87fe-3db7adf6cdab" (UID: "8c5385b7-40c0-4927-87fe-3db7adf6cdab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.244486 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-config-data" (OuterVolumeSpecName: "config-data") pod "8c5385b7-40c0-4927-87fe-3db7adf6cdab" (UID: "8c5385b7-40c0-4927-87fe-3db7adf6cdab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.245252 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d69c6e25-d85a-4ac9-ab43-17ee2752e5be" (UID: "d69c6e25-d85a-4ac9-ab43-17ee2752e5be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.251973 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f40b2a66-a5cd-4952-ac90-947bb0b46c28" (UID: "f40b2a66-a5cd-4952-ac90-947bb0b46c28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.252826 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0" (OuterVolumeSpecName: "glance") pod "d69c6e25-d85a-4ac9-ab43-17ee2752e5be" (UID: "d69c6e25-d85a-4ac9-ab43-17ee2752e5be"). InnerVolumeSpecName "pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.298433 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-config-data" (OuterVolumeSpecName: "config-data") pod "d69c6e25-d85a-4ac9-ab43-17ee2752e5be" (UID: "d69c6e25-d85a-4ac9-ab43-17ee2752e5be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.305893 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.305922 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.305934 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.305970 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") on node \"crc\" " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.305987 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.305998 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.306018 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") on node \"crc\" " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.306028 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp7pk\" (UniqueName: \"kubernetes.io/projected/8c5385b7-40c0-4927-87fe-3db7adf6cdab-kube-api-access-hp7pk\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.306038 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.306048 4744 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.306057 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.306066 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28rb6\" (UniqueName: \"kubernetes.io/projected/f40b2a66-a5cd-4952-ac90-947bb0b46c28-kube-api-access-28rb6\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.306077 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.306085 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28w4l\" (UniqueName: \"kubernetes.io/projected/d69c6e25-d85a-4ac9-ab43-17ee2752e5be-kube-api-access-28w4l\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.306093 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8c5385b7-40c0-4927-87fe-3db7adf6cdab-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.334263 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-config-data" (OuterVolumeSpecName: "config-data") pod "f40b2a66-a5cd-4952-ac90-947bb0b46c28" (UID: "f40b2a66-a5cd-4952-ac90-947bb0b46c28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.337077 4744 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.337292 4744 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0") on node "crc" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.338901 4744 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.339082 4744 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b") on node "crc" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.380675 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-tffgx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.380748 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-tffgx" event={"ID":"8c5385b7-40c0-4927-87fe-3db7adf6cdab","Type":"ContainerDied","Data":"daf852d856ecda8349f1971d99249fc85f011bcea2ceec16a4e85527881a8259"} Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.381114 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="daf852d856ecda8349f1971d99249fc85f011bcea2ceec16a4e85527881a8259" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.386315 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d69c6e25-d85a-4ac9-ab43-17ee2752e5be","Type":"ContainerDied","Data":"4fd51258a1d3082bcce99cd582769e15eab469a37abb443fdf8f56b8f59cd152"} Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.386394 4744 scope.go:117] "RemoveContainer" containerID="13782914df07095b238be4ab6d8eb719b7d53c1314b598f8b307e3c0b5879b3b" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.386612 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.420451 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f40b2a66-a5cd-4952-ac90-947bb0b46c28","Type":"ContainerDied","Data":"faf739dc7477888e7c38e12e6adc88fb5a197a45f77b140bcd45105c774dbc55"} Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.420536 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.423848 4744 reconciler_common.go:293] "Volume detached for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.424256 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f40b2a66-a5cd-4952-ac90-947bb0b46c28-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.424352 4744 reconciler_common.go:293] "Volume detached for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.484378 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.519209 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.533986 4744 scope.go:117] "RemoveContainer" containerID="032bad6126a09cc22dac483eebb2a7927f98b259f49ef8b33006a6e9d73613f5" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.538424 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-77f58bbbd5-m6wzx"] Jan 06 14:59:12 crc kubenswrapper[4744]: E0106 14:59:12.539114 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerName="glance-httpd" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539128 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerName="glance-httpd" Jan 06 14:59:12 crc kubenswrapper[4744]: E0106 14:59:12.539145 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerName="glance-log" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539151 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerName="glance-log" Jan 06 14:59:12 crc kubenswrapper[4744]: E0106 14:59:12.539183 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerName="glance-httpd" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539189 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerName="glance-httpd" Jan 06 14:59:12 crc kubenswrapper[4744]: E0106 14:59:12.539217 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c5385b7-40c0-4927-87fe-3db7adf6cdab" containerName="keystone-bootstrap" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539227 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c5385b7-40c0-4927-87fe-3db7adf6cdab" containerName="keystone-bootstrap" Jan 06 14:59:12 crc kubenswrapper[4744]: E0106 14:59:12.539234 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerName="glance-log" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539240 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerName="glance-log" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539433 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerName="glance-httpd" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539448 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" containerName="glance-log" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539462 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerName="glance-log" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539475 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c5385b7-40c0-4927-87fe-3db7adf6cdab" containerName="keystone-bootstrap" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.539483 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" containerName="glance-httpd" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.540246 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.543498 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.543636 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.543798 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.543902 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.544119 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8hqr4" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.544249 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.633080 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-77f58bbbd5-m6wzx"] Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.641515 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-combined-ca-bundle\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.641611 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-scripts\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.641637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-credential-keys\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.641657 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jzpx\" (UniqueName: \"kubernetes.io/projected/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-kube-api-access-2jzpx\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.641715 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-config-data\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.641764 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-fernet-keys\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.641780 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-public-tls-certs\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.641891 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-internal-tls-certs\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.662851 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.673405 4744 scope.go:117] "RemoveContainer" containerID="48cbf5a5016c23ffdaf2ff5a591e86f1648709cf93b8f2dc9a239fb778663832" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.712835 4744 scope.go:117] "RemoveContainer" containerID="44fbad7848baba929d60dee2bbd48d3085b46e976ece76ea3117345a243dc919" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.718532 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.727975 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.729920 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.729944 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.731815 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.733351 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.733630 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-lxm4v" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.735487 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.744962 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.749310 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-internal-tls-certs\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.749756 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-combined-ca-bundle\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.749914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-scripts\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.749953 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-credential-keys\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.749977 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jzpx\" (UniqueName: \"kubernetes.io/projected/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-kube-api-access-2jzpx\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.750095 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-config-data\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.750212 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-fernet-keys\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.750233 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-public-tls-certs\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.757501 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-fernet-keys\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.758752 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-public-tls-certs\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.759856 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-scripts\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.760715 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-internal-tls-certs\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.763600 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-credential-keys\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.764535 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-config-data\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.765698 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-combined-ca-bundle\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.768024 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 14:59:12 crc kubenswrapper[4744]: E0106 14:59:12.776810 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" containerName="dnsmasq-dns" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.776832 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" containerName="dnsmasq-dns" Jan 06 14:59:12 crc kubenswrapper[4744]: E0106 14:59:12.776849 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" containerName="init" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.776857 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" containerName="init" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.777073 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" containerName="dnsmasq-dns" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.778216 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.791730 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.793569 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jzpx\" (UniqueName: \"kubernetes.io/projected/6cee76ef-12c9-4949-aeeb-faea5a0f5ddb-kube-api-access-2jzpx\") pod \"keystone-77f58bbbd5-m6wzx\" (UID: \"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb\") " pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.803610 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.827671 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.851204 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-config\") pod \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.851308 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m65r\" (UniqueName: \"kubernetes.io/projected/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-kube-api-access-9m65r\") pod \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.851354 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-sb\") pod \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.851378 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-nb\") pod \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.851574 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-swift-storage-0\") pod \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.851597 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-svc\") pod \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\" (UID: \"987e985e-6a9f-4f21-8f8b-75ae7da8c45a\") " Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.851883 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.851940 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.851989 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.852031 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.852060 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.852140 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sztt\" (UniqueName: \"kubernetes.io/projected/992860bb-2626-401b-b4a0-439bb834f8e3-kube-api-access-4sztt\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.853317 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-logs\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.853433 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.863507 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-kube-api-access-9m65r" (OuterVolumeSpecName: "kube-api-access-9m65r") pod "987e985e-6a9f-4f21-8f8b-75ae7da8c45a" (UID: "987e985e-6a9f-4f21-8f8b-75ae7da8c45a"). InnerVolumeSpecName "kube-api-access-9m65r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.904061 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.925820 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "987e985e-6a9f-4f21-8f8b-75ae7da8c45a" (UID: "987e985e-6a9f-4f21-8f8b-75ae7da8c45a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.927645 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "987e985e-6a9f-4f21-8f8b-75ae7da8c45a" (UID: "987e985e-6a9f-4f21-8f8b-75ae7da8c45a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.940753 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-config" (OuterVolumeSpecName: "config") pod "987e985e-6a9f-4f21-8f8b-75ae7da8c45a" (UID: "987e985e-6a9f-4f21-8f8b-75ae7da8c45a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955597 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955640 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955722 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955749 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-logs\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955776 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955822 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955874 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955899 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955933 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955957 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl82v\" (UniqueName: \"kubernetes.io/projected/cab7dd39-8d15-462c-9df0-dc4df7e54db6-kube-api-access-zl82v\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.955974 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.956027 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sztt\" (UniqueName: \"kubernetes.io/projected/992860bb-2626-401b-b4a0-439bb834f8e3-kube-api-access-4sztt\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.956062 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.956084 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-logs\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.956135 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.956147 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m65r\" (UniqueName: \"kubernetes.io/projected/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-kube-api-access-9m65r\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.956172 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.956181 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.956621 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-logs\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.957377 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:12 crc kubenswrapper[4744]: I0106 14:59:12.982904 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "987e985e-6a9f-4f21-8f8b-75ae7da8c45a" (UID: "987e985e-6a9f-4f21-8f8b-75ae7da8c45a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.057225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.057258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.057323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.057352 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-logs\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.057453 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.057477 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl82v\" (UniqueName: \"kubernetes.io/projected/cab7dd39-8d15-462c-9df0-dc4df7e54db6-kube-api-access-zl82v\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.057497 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.057554 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.057611 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.058387 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-logs\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.058590 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.101787 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-config-data\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.106659 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "987e985e-6a9f-4f21-8f8b-75ae7da8c45a" (UID: "987e985e-6a9f-4f21-8f8b-75ae7da8c45a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.107428 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.107435 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.110332 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.110356 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.110617 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sztt\" (UniqueName: \"kubernetes.io/projected/992860bb-2626-401b-b4a0-439bb834f8e3-kube-api-access-4sztt\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.110726 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.111866 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-scripts\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.121907 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl82v\" (UniqueName: \"kubernetes.io/projected/cab7dd39-8d15-462c-9df0-dc4df7e54db6-kube-api-access-zl82v\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.122868 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.159958 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/987e985e-6a9f-4f21-8f8b-75ae7da8c45a-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.196206 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.196267 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a751ba72a1be54984524758ff1fa5b9a696d9d8d73f1feb0e198d80940b6b3a7/globalmount\"" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.196756 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.196849 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ac0041d382eb549e2b0d342e4bfe43b8e025e345c53521653e27358ce8689034/globalmount\"" pod="openstack/glance-default-external-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.483384 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.483424 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-xm2rp" event={"ID":"987e985e-6a9f-4f21-8f8b-75ae7da8c45a","Type":"ContainerDied","Data":"f3db661f67316563e570e94ba820ec4f5c1880f94ecac654463b38567f1cc598"} Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.484670 4744 scope.go:117] "RemoveContainer" containerID="7da164beb9c434748ceb41a8b419af03d092d18f8ec38a033f3931712201d946" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.530034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.537498 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " pod="openstack/glance-default-external-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.573623 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-xm2rp"] Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.589238 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-xm2rp"] Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.652517 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.711069 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.723266 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="987e985e-6a9f-4f21-8f8b-75ae7da8c45a" path="/var/lib/kubelet/pods/987e985e-6a9f-4f21-8f8b-75ae7da8c45a/volumes" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.723958 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d69c6e25-d85a-4ac9-ab43-17ee2752e5be" path="/var/lib/kubelet/pods/d69c6e25-d85a-4ac9-ab43-17ee2752e5be/volumes" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.724743 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f40b2a66-a5cd-4952-ac90-947bb0b46c28" path="/var/lib/kubelet/pods/f40b2a66-a5cd-4952-ac90-947bb0b46c28/volumes" Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.725905 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-77f58bbbd5-m6wzx"] Jan 06 14:59:13 crc kubenswrapper[4744]: I0106 14:59:13.753833 4744 scope.go:117] "RemoveContainer" containerID="9049aec0e90a0cefb223a825ee5e1006c1cfd39580006f567b4f1539c501e00b" Jan 06 14:59:14 crc kubenswrapper[4744]: I0106 14:59:14.588748 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-77f58bbbd5-m6wzx" event={"ID":"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb","Type":"ContainerStarted","Data":"7f64b2a1cb82371821c4adc248a64364b40701d9d4cfaf57a7d5106e0dcf6f9f"} Jan 06 14:59:15 crc kubenswrapper[4744]: I0106 14:59:15.128218 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 14:59:15 crc kubenswrapper[4744]: W0106 14:59:15.401663 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod992860bb_2626_401b_b4a0_439bb834f8e3.slice/crio-05de888a323b4e236b1dec49054d625def33bbb0f01d2c8fa315cf9128697c32 WatchSource:0}: Error finding container 05de888a323b4e236b1dec49054d625def33bbb0f01d2c8fa315cf9128697c32: Status 404 returned error can't find the container with id 05de888a323b4e236b1dec49054d625def33bbb0f01d2c8fa315cf9128697c32 Jan 06 14:59:15 crc kubenswrapper[4744]: I0106 14:59:15.403372 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 14:59:15 crc kubenswrapper[4744]: I0106 14:59:15.612995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"992860bb-2626-401b-b4a0-439bb834f8e3","Type":"ContainerStarted","Data":"05de888a323b4e236b1dec49054d625def33bbb0f01d2c8fa315cf9128697c32"} Jan 06 14:59:15 crc kubenswrapper[4744]: I0106 14:59:15.616035 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cab7dd39-8d15-462c-9df0-dc4df7e54db6","Type":"ContainerStarted","Data":"ea6a39d4640e2178dbdb96cef4a87983970f202703f7744008e8148e36e2d8d8"} Jan 06 14:59:16 crc kubenswrapper[4744]: I0106 14:59:16.645439 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-77f58bbbd5-m6wzx" event={"ID":"6cee76ef-12c9-4949-aeeb-faea5a0f5ddb","Type":"ContainerStarted","Data":"bfbf9fd684b5c3ff9771cc95e82d4d20570bbaf8635a2bc67890d19443f07134"} Jan 06 14:59:16 crc kubenswrapper[4744]: I0106 14:59:16.646136 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:16 crc kubenswrapper[4744]: I0106 14:59:16.656475 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cab7dd39-8d15-462c-9df0-dc4df7e54db6","Type":"ContainerStarted","Data":"0d55873db7667c2a58e81e8697f3b88667d3f505075d94fb72f3e82451ef5938"} Jan 06 14:59:16 crc kubenswrapper[4744]: I0106 14:59:16.695401 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-77f58bbbd5-m6wzx" podStartSLOduration=4.6953816889999995 podStartE2EDuration="4.695381689s" podCreationTimestamp="2026-01-06 14:59:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:59:16.686211207 +0000 UTC m=+1353.313677535" watchObservedRunningTime="2026-01-06 14:59:16.695381689 +0000 UTC m=+1353.322848007" Jan 06 14:59:17 crc kubenswrapper[4744]: I0106 14:59:17.670583 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"992860bb-2626-401b-b4a0-439bb834f8e3","Type":"ContainerStarted","Data":"f704408826be9ed92a9d2c0dfade364364a46376f5afb28dfca7b651afcd64f2"} Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.694481 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a","Type":"ContainerStarted","Data":"3d5a99b8ff824a4b736280ff646ba7fc1334587c1c2992e75d779289a8b04534"} Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.697448 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cab7dd39-8d15-462c-9df0-dc4df7e54db6","Type":"ContainerStarted","Data":"9281ea95ced4d12abbc3f13ffac89b4f12c76d9e8363006875383c1c108ad996"} Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.699367 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-m8c54" event={"ID":"4d50465b-09ea-4ff4-aa47-12f1c49ff207","Type":"ContainerStarted","Data":"ff7d2ade3aab7c9781162b8f5dbd880658f6b48bae8d4bef5e1ff547a738b44f"} Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.701973 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"992860bb-2626-401b-b4a0-439bb834f8e3","Type":"ContainerStarted","Data":"de4be668780cd37bacbeb7ad79259bca72d6ec15130dcb1a90d59bd1ee307017"} Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.704440 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vdnt5" event={"ID":"2b74efd3-9c21-4555-bfa8-5c1652b380dc","Type":"ContainerStarted","Data":"a4deba525591147b2da1e77d8a28191c995aff4a99f4ef3dd3ef23257f19c60a"} Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.708088 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbg7p" event={"ID":"e584b5b6-874e-48ea-b531-267a2c20562b","Type":"ContainerStarted","Data":"f4e9952d12c443bfe8f88057b4ecc3a8985e7e6b4e2e5adae25fff023a27ccaf"} Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.732618 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.732600828 podStartE2EDuration="7.732600828s" podCreationTimestamp="2026-01-06 14:59:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:59:19.719082981 +0000 UTC m=+1356.346549299" watchObservedRunningTime="2026-01-06 14:59:19.732600828 +0000 UTC m=+1356.360067146" Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.747688 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-vdnt5" podStartSLOduration=3.902461302 podStartE2EDuration="1m6.747672987s" podCreationTimestamp="2026-01-06 14:58:13 +0000 UTC" firstStartedPulling="2026-01-06 14:58:15.456619766 +0000 UTC m=+1292.084086084" lastFinishedPulling="2026-01-06 14:59:18.301831451 +0000 UTC m=+1354.929297769" observedRunningTime="2026-01-06 14:59:19.746188978 +0000 UTC m=+1356.373655306" watchObservedRunningTime="2026-01-06 14:59:19.747672987 +0000 UTC m=+1356.375139315" Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.770110 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.77009198 podStartE2EDuration="7.77009198s" podCreationTimestamp="2026-01-06 14:59:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:59:19.765012526 +0000 UTC m=+1356.392478864" watchObservedRunningTime="2026-01-06 14:59:19.77009198 +0000 UTC m=+1356.397558298" Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.791732 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-m8c54" podStartSLOduration=4.063266894 podStartE2EDuration="1m6.791716562s" podCreationTimestamp="2026-01-06 14:58:13 +0000 UTC" firstStartedPulling="2026-01-06 14:58:15.827768861 +0000 UTC m=+1292.455235189" lastFinishedPulling="2026-01-06 14:59:18.556218539 +0000 UTC m=+1355.183684857" observedRunningTime="2026-01-06 14:59:19.784425689 +0000 UTC m=+1356.411892027" watchObservedRunningTime="2026-01-06 14:59:19.791716562 +0000 UTC m=+1356.419182880" Jan 06 14:59:19 crc kubenswrapper[4744]: I0106 14:59:19.816021 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pbg7p" podStartSLOduration=4.843977461 podStartE2EDuration="1m8.816003354s" podCreationTimestamp="2026-01-06 14:58:11 +0000 UTC" firstStartedPulling="2026-01-06 14:58:14.30832175 +0000 UTC m=+1290.935788068" lastFinishedPulling="2026-01-06 14:59:18.280347643 +0000 UTC m=+1354.907813961" observedRunningTime="2026-01-06 14:59:19.808412523 +0000 UTC m=+1356.435878841" watchObservedRunningTime="2026-01-06 14:59:19.816003354 +0000 UTC m=+1356.443469672" Jan 06 14:59:21 crc kubenswrapper[4744]: I0106 14:59:21.407278 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:59:21 crc kubenswrapper[4744]: I0106 14:59:21.407950 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 14:59:21 crc kubenswrapper[4744]: I0106 14:59:21.742770 4744 generic.go:334] "Generic (PLEG): container finished" podID="be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" containerID="6dbf2f5d4c23d880674670d0d4a0482afb770c9fd6ec883381a24e1ed079b285" exitCode=0 Jan 06 14:59:21 crc kubenswrapper[4744]: I0106 14:59:21.742811 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dfxz9" event={"ID":"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b","Type":"ContainerDied","Data":"6dbf2f5d4c23d880674670d0d4a0482afb770c9fd6ec883381a24e1ed079b285"} Jan 06 14:59:22 crc kubenswrapper[4744]: I0106 14:59:22.470675 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pbg7p" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="registry-server" probeResult="failure" output=< Jan 06 14:59:22 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:59:22 crc kubenswrapper[4744]: > Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.653957 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.654323 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.692339 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.735685 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.735728 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.735854 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.756180 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.770186 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.772580 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.772622 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.772634 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 06 14:59:23 crc kubenswrapper[4744]: I0106 14:59:23.772644 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.281603 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dfxz9" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.421060 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-kube-api-access-5n6zx\") pod \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.421147 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-logs\") pod \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.421185 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-combined-ca-bundle\") pod \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.421236 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-config-data\") pod \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.421258 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-scripts\") pod \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\" (UID: \"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b\") " Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.421557 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-logs" (OuterVolumeSpecName: "logs") pod "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" (UID: "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.423264 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-logs\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.431515 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-scripts" (OuterVolumeSpecName: "scripts") pod "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" (UID: "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.431872 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-kube-api-access-5n6zx" (OuterVolumeSpecName: "kube-api-access-5n6zx") pod "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" (UID: "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b"). InnerVolumeSpecName "kube-api-access-5n6zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.453034 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-config-data" (OuterVolumeSpecName: "config-data") pod "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" (UID: "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.456363 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" (UID: "be9d6eb2-8ad6-4eb4-9208-a89adf751c4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.525245 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.525284 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.525294 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n6zx\" (UniqueName: \"kubernetes.io/projected/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-kube-api-access-5n6zx\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.525305 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.809857 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-dfxz9" event={"ID":"be9d6eb2-8ad6-4eb4-9208-a89adf751c4b","Type":"ContainerDied","Data":"b4ad122d47296d6847356b963f509d9d2d94a8744d46ea22e8e0d99d20fb22d0"} Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.810121 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4ad122d47296d6847356b963f509d9d2d94a8744d46ea22e8e0d99d20fb22d0" Jan 06 14:59:25 crc kubenswrapper[4744]: I0106 14:59:25.809937 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-dfxz9" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.451324 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-68984dbdcb-mkv9c"] Jan 06 14:59:26 crc kubenswrapper[4744]: E0106 14:59:26.451784 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" containerName="placement-db-sync" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.451799 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" containerName="placement-db-sync" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.452035 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" containerName="placement-db-sync" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.453212 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.457955 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.458289 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-m65j6" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.458415 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.458547 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.458847 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.471769 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68984dbdcb-mkv9c"] Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.560488 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-internal-tls-certs\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.560720 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-config-data\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.560889 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-public-tls-certs\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.561214 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-combined-ca-bundle\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.561250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-scripts\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.561280 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wmhb\" (UniqueName: \"kubernetes.io/projected/7d48c33e-540e-4bd3-b88a-592dff511604-kube-api-access-7wmhb\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.561364 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d48c33e-540e-4bd3-b88a-592dff511604-logs\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.672214 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d48c33e-540e-4bd3-b88a-592dff511604-logs\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.672338 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-internal-tls-certs\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.672479 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-config-data\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.672599 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-public-tls-certs\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.672648 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d48c33e-540e-4bd3-b88a-592dff511604-logs\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.672887 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-combined-ca-bundle\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.672929 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-scripts\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.672959 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wmhb\" (UniqueName: \"kubernetes.io/projected/7d48c33e-540e-4bd3-b88a-592dff511604-kube-api-access-7wmhb\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.677317 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-combined-ca-bundle\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.677345 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-public-tls-certs\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.677670 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-config-data\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.677964 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-internal-tls-certs\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.679061 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d48c33e-540e-4bd3-b88a-592dff511604-scripts\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.696558 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wmhb\" (UniqueName: \"kubernetes.io/projected/7d48c33e-540e-4bd3-b88a-592dff511604-kube-api-access-7wmhb\") pod \"placement-68984dbdcb-mkv9c\" (UID: \"7d48c33e-540e-4bd3-b88a-592dff511604\") " pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:26 crc kubenswrapper[4744]: I0106 14:59:26.770591 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:28 crc kubenswrapper[4744]: I0106 14:59:28.275834 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-68984dbdcb-mkv9c"] Jan 06 14:59:28 crc kubenswrapper[4744]: I0106 14:59:28.419955 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 06 14:59:28 crc kubenswrapper[4744]: I0106 14:59:28.420055 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 06 14:59:28 crc kubenswrapper[4744]: I0106 14:59:28.421176 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 06 14:59:28 crc kubenswrapper[4744]: I0106 14:59:28.421434 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:28 crc kubenswrapper[4744]: I0106 14:59:28.428529 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 06 14:59:29 crc kubenswrapper[4744]: W0106 14:59:29.173397 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d48c33e_540e_4bd3_b88a_592dff511604.slice/crio-eb7d0ce566dfd1add8820b38eec57421ad900b57b04bdf937ec4df050643aeca WatchSource:0}: Error finding container eb7d0ce566dfd1add8820b38eec57421ad900b57b04bdf937ec4df050643aeca: Status 404 returned error can't find the container with id eb7d0ce566dfd1add8820b38eec57421ad900b57b04bdf937ec4df050643aeca Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.621026 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d5sbl"] Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.623623 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.630921 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d5sbl"] Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.744787 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmj58\" (UniqueName: \"kubernetes.io/projected/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-kube-api-access-pmj58\") pod \"certified-operators-d5sbl\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.745117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-utilities\") pod \"certified-operators-d5sbl\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.745296 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-catalog-content\") pod \"certified-operators-d5sbl\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: E0106 14:59:29.815025 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.848251 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmj58\" (UniqueName: \"kubernetes.io/projected/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-kube-api-access-pmj58\") pod \"certified-operators-d5sbl\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.848302 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-utilities\") pod \"certified-operators-d5sbl\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.848325 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-catalog-content\") pod \"certified-operators-d5sbl\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.850910 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-catalog-content\") pod \"certified-operators-d5sbl\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.852332 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-utilities\") pod \"certified-operators-d5sbl\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.862423 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68984dbdcb-mkv9c" event={"ID":"7d48c33e-540e-4bd3-b88a-592dff511604","Type":"ContainerStarted","Data":"a6f551c4764e3b376270dad95ad03bf2272978867cf2c619f8c6543af4e5149a"} Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.862478 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68984dbdcb-mkv9c" event={"ID":"7d48c33e-540e-4bd3-b88a-592dff511604","Type":"ContainerStarted","Data":"eb7d0ce566dfd1add8820b38eec57421ad900b57b04bdf937ec4df050643aeca"} Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.868389 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmj58\" (UniqueName: \"kubernetes.io/projected/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-kube-api-access-pmj58\") pod \"certified-operators-d5sbl\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.903295 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a","Type":"ContainerStarted","Data":"fa05ef32e861ea6d3e39b13f193bf7b0ec2494d197c66e61e3920751429a755c"} Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.903533 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="ceilometer-notification-agent" containerID="cri-o://61047799c5a70e9d54f5a34f6084719581dd169ba38e6826d47527d3b52e0f41" gracePeriod=30 Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.903896 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.904310 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="proxy-httpd" containerID="cri-o://fa05ef32e861ea6d3e39b13f193bf7b0ec2494d197c66e61e3920751429a755c" gracePeriod=30 Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.904397 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="sg-core" containerID="cri-o://3d5a99b8ff824a4b736280ff646ba7fc1334587c1c2992e75d779289a8b04534" gracePeriod=30 Jan 06 14:59:29 crc kubenswrapper[4744]: I0106 14:59:29.941657 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:30 crc kubenswrapper[4744]: I0106 14:59:30.477079 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d5sbl"] Jan 06 14:59:30 crc kubenswrapper[4744]: I0106 14:59:30.915766 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5sbl" event={"ID":"b48f0e32-c65a-4d97-9006-4c43ce1f0b78","Type":"ContainerStarted","Data":"7017d1a4105a56b9f39afeab482bcc448f7dea891ebd47d182da04383017b5d3"} Jan 06 14:59:30 crc kubenswrapper[4744]: I0106 14:59:30.919195 4744 generic.go:334] "Generic (PLEG): container finished" podID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerID="3d5a99b8ff824a4b736280ff646ba7fc1334587c1c2992e75d779289a8b04534" exitCode=2 Jan 06 14:59:30 crc kubenswrapper[4744]: I0106 14:59:30.919248 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a","Type":"ContainerDied","Data":"3d5a99b8ff824a4b736280ff646ba7fc1334587c1c2992e75d779289a8b04534"} Jan 06 14:59:31 crc kubenswrapper[4744]: I0106 14:59:31.936860 4744 generic.go:334] "Generic (PLEG): container finished" podID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerID="fa05ef32e861ea6d3e39b13f193bf7b0ec2494d197c66e61e3920751429a755c" exitCode=0 Jan 06 14:59:31 crc kubenswrapper[4744]: I0106 14:59:31.937049 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a","Type":"ContainerDied","Data":"fa05ef32e861ea6d3e39b13f193bf7b0ec2494d197c66e61e3920751429a755c"} Jan 06 14:59:31 crc kubenswrapper[4744]: I0106 14:59:31.939570 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-68984dbdcb-mkv9c" event={"ID":"7d48c33e-540e-4bd3-b88a-592dff511604","Type":"ContainerStarted","Data":"8f391729bdea1fbecb78af15dabd0c26b1d7a650c89c3a38bc27f8f19c63ee58"} Jan 06 14:59:31 crc kubenswrapper[4744]: I0106 14:59:31.942084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5sbl" event={"ID":"b48f0e32-c65a-4d97-9006-4c43ce1f0b78","Type":"ContainerStarted","Data":"7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f"} Jan 06 14:59:32 crc kubenswrapper[4744]: I0106 14:59:32.454846 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pbg7p" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="registry-server" probeResult="failure" output=< Jan 06 14:59:32 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:59:32 crc kubenswrapper[4744]: > Jan 06 14:59:32 crc kubenswrapper[4744]: I0106 14:59:32.952877 4744 generic.go:334] "Generic (PLEG): container finished" podID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerID="7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f" exitCode=0 Jan 06 14:59:32 crc kubenswrapper[4744]: I0106 14:59:32.953001 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5sbl" event={"ID":"b48f0e32-c65a-4d97-9006-4c43ce1f0b78","Type":"ContainerDied","Data":"7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f"} Jan 06 14:59:32 crc kubenswrapper[4744]: I0106 14:59:32.953530 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:32 crc kubenswrapper[4744]: I0106 14:59:32.953594 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 14:59:32 crc kubenswrapper[4744]: I0106 14:59:32.999287 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-68984dbdcb-mkv9c" podStartSLOduration=6.999267565 podStartE2EDuration="6.999267565s" podCreationTimestamp="2026-01-06 14:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 14:59:32.991545161 +0000 UTC m=+1369.619011479" watchObservedRunningTime="2026-01-06 14:59:32.999267565 +0000 UTC m=+1369.626733893" Jan 06 14:59:35 crc kubenswrapper[4744]: I0106 14:59:35.991807 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5sbl" event={"ID":"b48f0e32-c65a-4d97-9006-4c43ce1f0b78","Type":"ContainerStarted","Data":"aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e"} Jan 06 14:59:35 crc kubenswrapper[4744]: I0106 14:59:35.994746 4744 generic.go:334] "Generic (PLEG): container finished" podID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerID="61047799c5a70e9d54f5a34f6084719581dd169ba38e6826d47527d3b52e0f41" exitCode=0 Jan 06 14:59:35 crc kubenswrapper[4744]: I0106 14:59:35.994790 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a","Type":"ContainerDied","Data":"61047799c5a70e9d54f5a34f6084719581dd169ba38e6826d47527d3b52e0f41"} Jan 06 14:59:37 crc kubenswrapper[4744]: I0106 14:59:37.009969 4744 generic.go:334] "Generic (PLEG): container finished" podID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerID="aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e" exitCode=0 Jan 06 14:59:37 crc kubenswrapper[4744]: I0106 14:59:37.010022 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5sbl" event={"ID":"b48f0e32-c65a-4d97-9006-4c43ce1f0b78","Type":"ContainerDied","Data":"aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e"} Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.205273 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.352716 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-sg-core-conf-yaml\") pod \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.352835 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-scripts\") pod \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.352862 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktcb5\" (UniqueName: \"kubernetes.io/projected/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-kube-api-access-ktcb5\") pod \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.352886 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-run-httpd\") pod \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.352926 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-config-data\") pod \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.353091 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-log-httpd\") pod \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.353197 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-combined-ca-bundle\") pod \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\" (UID: \"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a\") " Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.354376 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" (UID: "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.354414 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" (UID: "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.359257 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-kube-api-access-ktcb5" (OuterVolumeSpecName: "kube-api-access-ktcb5") pod "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" (UID: "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a"). InnerVolumeSpecName "kube-api-access-ktcb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.359536 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-scripts" (OuterVolumeSpecName: "scripts") pod "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" (UID: "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.416462 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" (UID: "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.443802 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" (UID: "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.456421 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.456449 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.456462 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.456471 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.456479 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktcb5\" (UniqueName: \"kubernetes.io/projected/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-kube-api-access-ktcb5\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.456488 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.479404 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-config-data" (OuterVolumeSpecName: "config-data") pod "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" (UID: "0832cc9a-5d34-4e11-8d3b-cc7c0674b73a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:38 crc kubenswrapper[4744]: I0106 14:59:38.559890 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.040511 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0832cc9a-5d34-4e11-8d3b-cc7c0674b73a","Type":"ContainerDied","Data":"f1ecd45a43453fcf957dc964d93a7f7d299de053fc5b0a834c1733630a2f3e1f"} Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.040586 4744 scope.go:117] "RemoveContainer" containerID="fa05ef32e861ea6d3e39b13f193bf7b0ec2494d197c66e61e3920751429a755c" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.040635 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.068892 4744 scope.go:117] "RemoveContainer" containerID="3d5a99b8ff824a4b736280ff646ba7fc1334587c1c2992e75d779289a8b04534" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.108868 4744 scope.go:117] "RemoveContainer" containerID="61047799c5a70e9d54f5a34f6084719581dd169ba38e6826d47527d3b52e0f41" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.155871 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.169131 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.179697 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 14:59:39 crc kubenswrapper[4744]: E0106 14:59:39.180315 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="sg-core" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.180340 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="sg-core" Jan 06 14:59:39 crc kubenswrapper[4744]: E0106 14:59:39.180361 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="proxy-httpd" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.180368 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="proxy-httpd" Jan 06 14:59:39 crc kubenswrapper[4744]: E0106 14:59:39.180388 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="ceilometer-notification-agent" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.180394 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="ceilometer-notification-agent" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.180591 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="proxy-httpd" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.180612 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="ceilometer-notification-agent" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.180632 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" containerName="sg-core" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.183227 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.185187 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.185730 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.189499 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.278946 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-config-data\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.279057 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-log-httpd\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.279108 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhfkr\" (UniqueName: \"kubernetes.io/projected/c7966af7-509d-4b51-aa77-35a21616df45-kube-api-access-jhfkr\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.279209 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.279361 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-scripts\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.279469 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-run-httpd\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.279518 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.381258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhfkr\" (UniqueName: \"kubernetes.io/projected/c7966af7-509d-4b51-aa77-35a21616df45-kube-api-access-jhfkr\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.381370 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.381504 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-scripts\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.381634 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-run-httpd\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.381679 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.381912 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-config-data\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.381986 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-log-httpd\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.382555 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-log-httpd\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.382545 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-run-httpd\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.385795 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.386115 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-scripts\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.387376 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.394476 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-config-data\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.408256 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhfkr\" (UniqueName: \"kubernetes.io/projected/c7966af7-509d-4b51-aa77-35a21616df45-kube-api-access-jhfkr\") pod \"ceilometer-0\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.501121 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 14:59:39 crc kubenswrapper[4744]: I0106 14:59:39.726481 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0832cc9a-5d34-4e11-8d3b-cc7c0674b73a" path="/var/lib/kubelet/pods/0832cc9a-5d34-4e11-8d3b-cc7c0674b73a/volumes" Jan 06 14:59:40 crc kubenswrapper[4744]: I0106 14:59:40.006402 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 14:59:40 crc kubenswrapper[4744]: I0106 14:59:40.060682 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerStarted","Data":"99fad64f1b9c5778596c0d95f19d13f49f060be6f388c723a1e718cb4c329f4c"} Jan 06 14:59:42 crc kubenswrapper[4744]: I0106 14:59:42.491234 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pbg7p" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="registry-server" probeResult="failure" output=< Jan 06 14:59:42 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:59:42 crc kubenswrapper[4744]: > Jan 06 14:59:45 crc kubenswrapper[4744]: I0106 14:59:45.136827 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5sbl" event={"ID":"b48f0e32-c65a-4d97-9006-4c43ce1f0b78","Type":"ContainerStarted","Data":"9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6"} Jan 06 14:59:46 crc kubenswrapper[4744]: I0106 14:59:46.181727 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d5sbl" podStartSLOduration=5.629699051 podStartE2EDuration="17.181700064s" podCreationTimestamp="2026-01-06 14:59:29 +0000 UTC" firstStartedPulling="2026-01-06 14:59:32.954531162 +0000 UTC m=+1369.581997480" lastFinishedPulling="2026-01-06 14:59:44.506532155 +0000 UTC m=+1381.133998493" observedRunningTime="2026-01-06 14:59:46.17432147 +0000 UTC m=+1382.801787788" watchObservedRunningTime="2026-01-06 14:59:46.181700064 +0000 UTC m=+1382.809166372" Jan 06 14:59:46 crc kubenswrapper[4744]: I0106 14:59:46.298180 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-77f58bbbd5-m6wzx" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.377882 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.380044 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.385595 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.385674 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.385854 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-cpk2p" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.408291 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.524588 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.524709 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.524842 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config-secret\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.525098 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjmdx\" (UniqueName: \"kubernetes.io/projected/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-kube-api-access-vjmdx\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.627554 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.627631 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config-secret\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.627697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjmdx\" (UniqueName: \"kubernetes.io/projected/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-kube-api-access-vjmdx\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.627761 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.628527 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.636078 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.650269 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config-secret\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.664908 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjmdx\" (UniqueName: \"kubernetes.io/projected/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-kube-api-access-vjmdx\") pod \"openstackclient\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.708746 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.823583 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.857423 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.868695 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.870074 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 06 14:59:48 crc kubenswrapper[4744]: I0106 14:59:48.898656 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.039120 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50332b7c-a1e8-4cd4-95d6-463173669507-combined-ca-bundle\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.039194 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/50332b7c-a1e8-4cd4-95d6-463173669507-openstack-config-secret\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.039412 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgcxc\" (UniqueName: \"kubernetes.io/projected/50332b7c-a1e8-4cd4-95d6-463173669507-kube-api-access-sgcxc\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.039477 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/50332b7c-a1e8-4cd4-95d6-463173669507-openstack-config\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.141382 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50332b7c-a1e8-4cd4-95d6-463173669507-combined-ca-bundle\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.141445 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/50332b7c-a1e8-4cd4-95d6-463173669507-openstack-config-secret\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.141519 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgcxc\" (UniqueName: \"kubernetes.io/projected/50332b7c-a1e8-4cd4-95d6-463173669507-kube-api-access-sgcxc\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.141537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/50332b7c-a1e8-4cd4-95d6-463173669507-openstack-config\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.142707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/50332b7c-a1e8-4cd4-95d6-463173669507-openstack-config\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.148687 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/50332b7c-a1e8-4cd4-95d6-463173669507-openstack-config-secret\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.149707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50332b7c-a1e8-4cd4-95d6-463173669507-combined-ca-bundle\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.190703 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgcxc\" (UniqueName: \"kubernetes.io/projected/50332b7c-a1e8-4cd4-95d6-463173669507-kube-api-access-sgcxc\") pod \"openstackclient\" (UID: \"50332b7c-a1e8-4cd4-95d6-463173669507\") " pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.193179 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.196488 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerStarted","Data":"c9e09149319fbba6a28036746c6d5bddf814c68e2b770403b877daccfcd20920"} Jan 06 14:59:49 crc kubenswrapper[4744]: E0106 14:59:49.305344 4744 log.go:32] "RunPodSandbox from runtime service failed" err=< Jan 06 14:59:49 crc kubenswrapper[4744]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_ef9b4d9d-05b5-47c7-b09f-d7f85d824361_0(6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467" Netns:"/var/run/netns/118545ea-6dc5-417f-8fc5-9e90fd605ccc" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467;K8S_POD_UID=ef9b4d9d-05b5-47c7-b09f-d7f85d824361" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: [openstack/openstackclient/ef9b4d9d-05b5-47c7-b09f-d7f85d824361:ovn-kubernetes]: error adding container to network "ovn-kubernetes": CNI request failed with status 400: '[openstack/openstackclient 6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467 network default NAD default] [openstack/openstackclient 6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467 network default NAD default] pod deleted before sandbox ADD operation began Jan 06 14:59:49 crc kubenswrapper[4744]: ' Jan 06 14:59:49 crc kubenswrapper[4744]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 06 14:59:49 crc kubenswrapper[4744]: > Jan 06 14:59:49 crc kubenswrapper[4744]: E0106 14:59:49.305658 4744 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Jan 06 14:59:49 crc kubenswrapper[4744]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_ef9b4d9d-05b5-47c7-b09f-d7f85d824361_0(6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467" Netns:"/var/run/netns/118545ea-6dc5-417f-8fc5-9e90fd605ccc" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467;K8S_POD_UID=ef9b4d9d-05b5-47c7-b09f-d7f85d824361" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: [openstack/openstackclient/ef9b4d9d-05b5-47c7-b09f-d7f85d824361:ovn-kubernetes]: error adding container to network "ovn-kubernetes": CNI request failed with status 400: '[openstack/openstackclient 6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467 network default NAD default] [openstack/openstackclient 6bada4f45dedc07de52eeb87391156977d7dc98022420b13e204f7fd82532467 network default NAD default] pod deleted before sandbox ADD operation began Jan 06 14:59:49 crc kubenswrapper[4744]: ' Jan 06 14:59:49 crc kubenswrapper[4744]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Jan 06 14:59:49 crc kubenswrapper[4744]: > pod="openstack/openstackclient" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.707207 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.942390 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:49 crc kubenswrapper[4744]: I0106 14:59:49.942648 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.208000 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.207988 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"50332b7c-a1e8-4cd4-95d6-463173669507","Type":"ContainerStarted","Data":"efedd2442c831e457687bf994f55131c7dda0fa11276a4cea9ac299512731f0d"} Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.221708 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.225732 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ef9b4d9d-05b5-47c7-b09f-d7f85d824361" podUID="50332b7c-a1e8-4cd4-95d6-463173669507" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.368533 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config\") pod \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.368614 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config-secret\") pod \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.368746 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjmdx\" (UniqueName: \"kubernetes.io/projected/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-kube-api-access-vjmdx\") pod \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.368765 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-combined-ca-bundle\") pod \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\" (UID: \"ef9b4d9d-05b5-47c7-b09f-d7f85d824361\") " Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.369148 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ef9b4d9d-05b5-47c7-b09f-d7f85d824361" (UID: "ef9b4d9d-05b5-47c7-b09f-d7f85d824361"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.369398 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.372539 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ef9b4d9d-05b5-47c7-b09f-d7f85d824361" (UID: "ef9b4d9d-05b5-47c7-b09f-d7f85d824361"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.374806 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef9b4d9d-05b5-47c7-b09f-d7f85d824361" (UID: "ef9b4d9d-05b5-47c7-b09f-d7f85d824361"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.375894 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-kube-api-access-vjmdx" (OuterVolumeSpecName: "kube-api-access-vjmdx") pod "ef9b4d9d-05b5-47c7-b09f-d7f85d824361" (UID: "ef9b4d9d-05b5-47c7-b09f-d7f85d824361"). InnerVolumeSpecName "kube-api-access-vjmdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.471997 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.472470 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjmdx\" (UniqueName: \"kubernetes.io/projected/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-kube-api-access-vjmdx\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:50 crc kubenswrapper[4744]: I0106 14:59:50.472487 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef9b4d9d-05b5-47c7-b09f-d7f85d824361-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 14:59:51 crc kubenswrapper[4744]: I0106 14:59:51.008538 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-d5sbl" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerName="registry-server" probeResult="failure" output=< Jan 06 14:59:51 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:59:51 crc kubenswrapper[4744]: > Jan 06 14:59:51 crc kubenswrapper[4744]: I0106 14:59:51.240329 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 06 14:59:51 crc kubenswrapper[4744]: I0106 14:59:51.261257 4744 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ef9b4d9d-05b5-47c7-b09f-d7f85d824361" podUID="50332b7c-a1e8-4cd4-95d6-463173669507" Jan 06 14:59:51 crc kubenswrapper[4744]: I0106 14:59:51.727662 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef9b4d9d-05b5-47c7-b09f-d7f85d824361" path="/var/lib/kubelet/pods/ef9b4d9d-05b5-47c7-b09f-d7f85d824361/volumes" Jan 06 14:59:52 crc kubenswrapper[4744]: I0106 14:59:52.253021 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerStarted","Data":"fe80bcca808b45ed0e1be2e4996875a0a96b62f5701c5854325766fa27f720af"} Jan 06 14:59:52 crc kubenswrapper[4744]: I0106 14:59:52.454813 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pbg7p" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="registry-server" probeResult="failure" output=< Jan 06 14:59:52 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 14:59:52 crc kubenswrapper[4744]: > Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.004407 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.055346 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.137985 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2"] Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.139648 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.142123 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.142526 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.166350 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2"] Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.249759 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa2592b1-dd21-4676-bea3-d032eb197ebb-secret-volume\") pod \"collect-profiles-29461860-rshx2\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.249802 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa2592b1-dd21-4676-bea3-d032eb197ebb-config-volume\") pod \"collect-profiles-29461860-rshx2\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.249896 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcrgt\" (UniqueName: \"kubernetes.io/projected/aa2592b1-dd21-4676-bea3-d032eb197ebb-kube-api-access-kcrgt\") pod \"collect-profiles-29461860-rshx2\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.351137 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa2592b1-dd21-4676-bea3-d032eb197ebb-secret-volume\") pod \"collect-profiles-29461860-rshx2\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.351207 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa2592b1-dd21-4676-bea3-d032eb197ebb-config-volume\") pod \"collect-profiles-29461860-rshx2\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.351278 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcrgt\" (UniqueName: \"kubernetes.io/projected/aa2592b1-dd21-4676-bea3-d032eb197ebb-kube-api-access-kcrgt\") pod \"collect-profiles-29461860-rshx2\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.352230 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa2592b1-dd21-4676-bea3-d032eb197ebb-config-volume\") pod \"collect-profiles-29461860-rshx2\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.357303 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa2592b1-dd21-4676-bea3-d032eb197ebb-secret-volume\") pod \"collect-profiles-29461860-rshx2\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.369537 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcrgt\" (UniqueName: \"kubernetes.io/projected/aa2592b1-dd21-4676-bea3-d032eb197ebb-kube-api-access-kcrgt\") pod \"collect-profiles-29461860-rshx2\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.479986 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:00 crc kubenswrapper[4744]: I0106 15:00:00.813340 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d5sbl"] Jan 06 15:00:01 crc kubenswrapper[4744]: I0106 15:00:01.358223 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d5sbl" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerName="registry-server" containerID="cri-o://9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6" gracePeriod=2 Jan 06 15:00:01 crc kubenswrapper[4744]: I0106 15:00:01.464420 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 15:00:01 crc kubenswrapper[4744]: I0106 15:00:01.586626 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.052251 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 15:00:02 crc kubenswrapper[4744]: W0106 15:00:02.165667 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa2592b1_dd21_4676_bea3_d032eb197ebb.slice/crio-554389cc3fba9b1818d65666dd4c9288b3cf39dae135e9e5f9da2acad0eb8e0b WatchSource:0}: Error finding container 554389cc3fba9b1818d65666dd4c9288b3cf39dae135e9e5f9da2acad0eb8e0b: Status 404 returned error can't find the container with id 554389cc3fba9b1818d65666dd4c9288b3cf39dae135e9e5f9da2acad0eb8e0b Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.167644 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2"] Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.204635 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-catalog-content\") pod \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.204745 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-utilities\") pod \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.204796 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmj58\" (UniqueName: \"kubernetes.io/projected/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-kube-api-access-pmj58\") pod \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\" (UID: \"b48f0e32-c65a-4d97-9006-4c43ce1f0b78\") " Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.205485 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-utilities" (OuterVolumeSpecName: "utilities") pod "b48f0e32-c65a-4d97-9006-4c43ce1f0b78" (UID: "b48f0e32-c65a-4d97-9006-4c43ce1f0b78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.210764 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-kube-api-access-pmj58" (OuterVolumeSpecName: "kube-api-access-pmj58") pod "b48f0e32-c65a-4d97-9006-4c43ce1f0b78" (UID: "b48f0e32-c65a-4d97-9006-4c43ce1f0b78"). InnerVolumeSpecName "kube-api-access-pmj58". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.257646 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b48f0e32-c65a-4d97-9006-4c43ce1f0b78" (UID: "b48f0e32-c65a-4d97-9006-4c43ce1f0b78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.307397 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmj58\" (UniqueName: \"kubernetes.io/projected/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-kube-api-access-pmj58\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.307484 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.307505 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b48f0e32-c65a-4d97-9006-4c43ce1f0b78-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.369957 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" event={"ID":"aa2592b1-dd21-4676-bea3-d032eb197ebb","Type":"ContainerStarted","Data":"554389cc3fba9b1818d65666dd4c9288b3cf39dae135e9e5f9da2acad0eb8e0b"} Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.372771 4744 generic.go:334] "Generic (PLEG): container finished" podID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerID="9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6" exitCode=0 Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.372809 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5sbl" event={"ID":"b48f0e32-c65a-4d97-9006-4c43ce1f0b78","Type":"ContainerDied","Data":"9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6"} Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.372845 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d5sbl" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.372860 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d5sbl" event={"ID":"b48f0e32-c65a-4d97-9006-4c43ce1f0b78","Type":"ContainerDied","Data":"7017d1a4105a56b9f39afeab482bcc448f7dea891ebd47d182da04383017b5d3"} Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.372887 4744 scope.go:117] "RemoveContainer" containerID="9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.407064 4744 scope.go:117] "RemoveContainer" containerID="aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.423575 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d5sbl"] Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.439030 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d5sbl"] Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.481881 4744 scope.go:117] "RemoveContainer" containerID="7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.514432 4744 scope.go:117] "RemoveContainer" containerID="9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6" Jan 06 15:00:02 crc kubenswrapper[4744]: E0106 15:00:02.516227 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6\": container with ID starting with 9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6 not found: ID does not exist" containerID="9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.516261 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6"} err="failed to get container status \"9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6\": rpc error: code = NotFound desc = could not find container \"9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6\": container with ID starting with 9f242e7f1fb00209330b6e8aaf93c052ae6696113a71d1ef1544e4e4da2934b6 not found: ID does not exist" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.516289 4744 scope.go:117] "RemoveContainer" containerID="aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e" Jan 06 15:00:02 crc kubenswrapper[4744]: E0106 15:00:02.516659 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e\": container with ID starting with aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e not found: ID does not exist" containerID="aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.516690 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e"} err="failed to get container status \"aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e\": rpc error: code = NotFound desc = could not find container \"aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e\": container with ID starting with aea94005d2f94575c05e1256ffd0b60ab8ac4285fe3353836fd53f6af671af1e not found: ID does not exist" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.516709 4744 scope.go:117] "RemoveContainer" containerID="7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f" Jan 06 15:00:02 crc kubenswrapper[4744]: E0106 15:00:02.517632 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f\": container with ID starting with 7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f not found: ID does not exist" containerID="7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.517665 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f"} err="failed to get container status \"7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f\": rpc error: code = NotFound desc = could not find container \"7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f\": container with ID starting with 7fce55c07e6405ebb0d6f662abc1e3636d111d091b37c90cb5ac62826f45660f not found: ID does not exist" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.554953 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 15:00:02 crc kubenswrapper[4744]: I0106 15:00:02.560853 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-68984dbdcb-mkv9c" Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.386515 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"50332b7c-a1e8-4cd4-95d6-463173669507","Type":"ContainerStarted","Data":"44eb849142a2d187b2d5e4bf323f414a15c529b5f2e79254f6e74f2eff333f73"} Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.388325 4744 generic.go:334] "Generic (PLEG): container finished" podID="4d50465b-09ea-4ff4-aa47-12f1c49ff207" containerID="ff7d2ade3aab7c9781162b8f5dbd880658f6b48bae8d4bef5e1ff547a738b44f" exitCode=0 Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.388382 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-m8c54" event={"ID":"4d50465b-09ea-4ff4-aa47-12f1c49ff207","Type":"ContainerDied","Data":"ff7d2ade3aab7c9781162b8f5dbd880658f6b48bae8d4bef5e1ff547a738b44f"} Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.392795 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerStarted","Data":"bdd65b64eac259914eb01ea6b6212e354d572e83d74fb6723c6b5db93e2ead7f"} Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.397920 4744 generic.go:334] "Generic (PLEG): container finished" podID="aa2592b1-dd21-4676-bea3-d032eb197ebb" containerID="f6de9cf249d4b29fe4a377e8df6f3754b2ded2846bab01bb5deb176ad54c09a1" exitCode=0 Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.398007 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" event={"ID":"aa2592b1-dd21-4676-bea3-d032eb197ebb","Type":"ContainerDied","Data":"f6de9cf249d4b29fe4a377e8df6f3754b2ded2846bab01bb5deb176ad54c09a1"} Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.407048 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.602882772 podStartE2EDuration="15.407028737s" podCreationTimestamp="2026-01-06 14:59:48 +0000 UTC" firstStartedPulling="2026-01-06 14:59:49.710476303 +0000 UTC m=+1386.337942661" lastFinishedPulling="2026-01-06 15:00:02.514622308 +0000 UTC m=+1399.142088626" observedRunningTime="2026-01-06 15:00:03.401631965 +0000 UTC m=+1400.029098283" watchObservedRunningTime="2026-01-06 15:00:03.407028737 +0000 UTC m=+1400.034495055" Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.732605 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" path="/var/lib/kubelet/pods/b48f0e32-c65a-4d97-9006-4c43ce1f0b78/volumes" Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.818399 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pbg7p"] Jan 06 15:00:03 crc kubenswrapper[4744]: I0106 15:00:03.818858 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pbg7p" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="registry-server" containerID="cri-o://f4e9952d12c443bfe8f88057b4ecc3a8985e7e6b4e2e5adae25fff023a27ccaf" gracePeriod=2 Jan 06 15:00:04 crc kubenswrapper[4744]: I0106 15:00:04.901947 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-m8c54" Jan 06 15:00:04 crc kubenswrapper[4744]: I0106 15:00:04.909460 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.082738 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-db-sync-config-data\") pod \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.082836 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa2592b1-dd21-4676-bea3-d032eb197ebb-config-volume\") pod \"aa2592b1-dd21-4676-bea3-d032eb197ebb\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.082912 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcrgt\" (UniqueName: \"kubernetes.io/projected/aa2592b1-dd21-4676-bea3-d032eb197ebb-kube-api-access-kcrgt\") pod \"aa2592b1-dd21-4676-bea3-d032eb197ebb\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.082984 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-combined-ca-bundle\") pod \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.083045 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa2592b1-dd21-4676-bea3-d032eb197ebb-secret-volume\") pod \"aa2592b1-dd21-4676-bea3-d032eb197ebb\" (UID: \"aa2592b1-dd21-4676-bea3-d032eb197ebb\") " Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.083068 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gbb6\" (UniqueName: \"kubernetes.io/projected/4d50465b-09ea-4ff4-aa47-12f1c49ff207-kube-api-access-7gbb6\") pod \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\" (UID: \"4d50465b-09ea-4ff4-aa47-12f1c49ff207\") " Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.085207 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa2592b1-dd21-4676-bea3-d032eb197ebb-config-volume" (OuterVolumeSpecName: "config-volume") pod "aa2592b1-dd21-4676-bea3-d032eb197ebb" (UID: "aa2592b1-dd21-4676-bea3-d032eb197ebb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.089530 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d50465b-09ea-4ff4-aa47-12f1c49ff207-kube-api-access-7gbb6" (OuterVolumeSpecName: "kube-api-access-7gbb6") pod "4d50465b-09ea-4ff4-aa47-12f1c49ff207" (UID: "4d50465b-09ea-4ff4-aa47-12f1c49ff207"). InnerVolumeSpecName "kube-api-access-7gbb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.090340 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa2592b1-dd21-4676-bea3-d032eb197ebb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "aa2592b1-dd21-4676-bea3-d032eb197ebb" (UID: "aa2592b1-dd21-4676-bea3-d032eb197ebb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.090729 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa2592b1-dd21-4676-bea3-d032eb197ebb-kube-api-access-kcrgt" (OuterVolumeSpecName: "kube-api-access-kcrgt") pod "aa2592b1-dd21-4676-bea3-d032eb197ebb" (UID: "aa2592b1-dd21-4676-bea3-d032eb197ebb"). InnerVolumeSpecName "kube-api-access-kcrgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.094568 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4d50465b-09ea-4ff4-aa47-12f1c49ff207" (UID: "4d50465b-09ea-4ff4-aa47-12f1c49ff207"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.136340 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d50465b-09ea-4ff4-aa47-12f1c49ff207" (UID: "4d50465b-09ea-4ff4-aa47-12f1c49ff207"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.187870 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa2592b1-dd21-4676-bea3-d032eb197ebb-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.187925 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gbb6\" (UniqueName: \"kubernetes.io/projected/4d50465b-09ea-4ff4-aa47-12f1c49ff207-kube-api-access-7gbb6\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.187957 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.187973 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa2592b1-dd21-4676-bea3-d032eb197ebb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.187993 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcrgt\" (UniqueName: \"kubernetes.io/projected/aa2592b1-dd21-4676-bea3-d032eb197ebb-kube-api-access-kcrgt\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.188009 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d50465b-09ea-4ff4-aa47-12f1c49ff207-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.425146 4744 generic.go:334] "Generic (PLEG): container finished" podID="e584b5b6-874e-48ea-b531-267a2c20562b" containerID="f4e9952d12c443bfe8f88057b4ecc3a8985e7e6b4e2e5adae25fff023a27ccaf" exitCode=0 Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.425475 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbg7p" event={"ID":"e584b5b6-874e-48ea-b531-267a2c20562b","Type":"ContainerDied","Data":"f4e9952d12c443bfe8f88057b4ecc3a8985e7e6b4e2e5adae25fff023a27ccaf"} Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.427430 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" event={"ID":"aa2592b1-dd21-4676-bea3-d032eb197ebb","Type":"ContainerDied","Data":"554389cc3fba9b1818d65666dd4c9288b3cf39dae135e9e5f9da2acad0eb8e0b"} Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.427504 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="554389cc3fba9b1818d65666dd4c9288b3cf39dae135e9e5f9da2acad0eb8e0b" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.427455 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.429211 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-m8c54" event={"ID":"4d50465b-09ea-4ff4-aa47-12f1c49ff207","Type":"ContainerDied","Data":"10eb4cd86fc5b9c79d571ad70fd8f05c629458bd9f01ac45f97feefc587d4607"} Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.429260 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10eb4cd86fc5b9c79d571ad70fd8f05c629458bd9f01ac45f97feefc587d4607" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.429305 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-m8c54" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.762301 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-74dd54d475-wqvpn"] Jan 06 15:00:05 crc kubenswrapper[4744]: E0106 15:00:05.762939 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerName="registry-server" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.762956 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerName="registry-server" Jan 06 15:00:05 crc kubenswrapper[4744]: E0106 15:00:05.762975 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d50465b-09ea-4ff4-aa47-12f1c49ff207" containerName="barbican-db-sync" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.762982 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d50465b-09ea-4ff4-aa47-12f1c49ff207" containerName="barbican-db-sync" Jan 06 15:00:05 crc kubenswrapper[4744]: E0106 15:00:05.762992 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerName="extract-content" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.762998 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerName="extract-content" Jan 06 15:00:05 crc kubenswrapper[4744]: E0106 15:00:05.763018 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerName="extract-utilities" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.763024 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerName="extract-utilities" Jan 06 15:00:05 crc kubenswrapper[4744]: E0106 15:00:05.763033 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa2592b1-dd21-4676-bea3-d032eb197ebb" containerName="collect-profiles" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.763039 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa2592b1-dd21-4676-bea3-d032eb197ebb" containerName="collect-profiles" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.763239 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d50465b-09ea-4ff4-aa47-12f1c49ff207" containerName="barbican-db-sync" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.763256 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b48f0e32-c65a-4d97-9006-4c43ce1f0b78" containerName="registry-server" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.763276 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa2592b1-dd21-4676-bea3-d032eb197ebb" containerName="collect-profiles" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.764405 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.770176 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.775542 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kzz6d" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.791401 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.825471 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-74dd54d475-wqvpn"] Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.888217 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5d48fc6d-cr8dv"] Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.890085 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.898475 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.919997 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc804009-f9fd-4d70-bdd8-c6c61a544549-combined-ca-bundle\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.920122 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc804009-f9fd-4d70-bdd8-c6c61a544549-config-data-custom\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.921211 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc804009-f9fd-4d70-bdd8-c6c61a544549-logs\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.921248 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zctw\" (UniqueName: \"kubernetes.io/projected/bc804009-f9fd-4d70-bdd8-c6c61a544549-kube-api-access-5zctw\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.921423 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc804009-f9fd-4d70-bdd8-c6c61a544549-config-data\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.932800 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5d48fc6d-cr8dv"] Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.950797 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-m48b5"] Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.953577 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:05 crc kubenswrapper[4744]: I0106 15:00:05.969497 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-m48b5"] Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024564 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-logs\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024654 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-combined-ca-bundle\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024684 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc804009-f9fd-4d70-bdd8-c6c61a544549-config-data-custom\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024722 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc804009-f9fd-4d70-bdd8-c6c61a544549-logs\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024744 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zctw\" (UniqueName: \"kubernetes.io/projected/bc804009-f9fd-4d70-bdd8-c6c61a544549-kube-api-access-5zctw\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024785 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-config-data\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024820 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-config-data-custom\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024836 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc804009-f9fd-4d70-bdd8-c6c61a544549-config-data\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024852 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bnfj\" (UniqueName: \"kubernetes.io/projected/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-kube-api-access-8bnfj\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.024900 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc804009-f9fd-4d70-bdd8-c6c61a544549-combined-ca-bundle\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.025793 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc804009-f9fd-4d70-bdd8-c6c61a544549-logs\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.033874 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bc804009-f9fd-4d70-bdd8-c6c61a544549-config-data-custom\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.043105 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc804009-f9fd-4d70-bdd8-c6c61a544549-combined-ca-bundle\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.051354 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zctw\" (UniqueName: \"kubernetes.io/projected/bc804009-f9fd-4d70-bdd8-c6c61a544549-kube-api-access-5zctw\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.057335 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc804009-f9fd-4d70-bdd8-c6c61a544549-config-data\") pod \"barbican-worker-74dd54d475-wqvpn\" (UID: \"bc804009-f9fd-4d70-bdd8-c6c61a544549\") " pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.063222 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7bc78458f8-mx59q"] Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.065299 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.071811 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.076001 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bc78458f8-mx59q"] Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.095782 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-74dd54d475-wqvpn" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135407 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-logs\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135554 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-combined-ca-bundle\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135576 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135645 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135669 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-config\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr2gj\" (UniqueName: \"kubernetes.io/projected/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-kube-api-access-wr2gj\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135724 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135749 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-config-data\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135804 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-config-data-custom\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.135822 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bnfj\" (UniqueName: \"kubernetes.io/projected/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-kube-api-access-8bnfj\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.136620 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-logs\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.142098 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-combined-ca-bundle\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.145020 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-config-data\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.147643 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-config-data-custom\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.158432 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bnfj\" (UniqueName: \"kubernetes.io/projected/762cdea3-aba0-4f8f-96c8-4dfaab8bf92a-kube-api-access-8bnfj\") pod \"barbican-keystone-listener-5d48fc6d-cr8dv\" (UID: \"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a\") " pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.223057 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.237953 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238024 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data-custom\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238403 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn6cb\" (UniqueName: \"kubernetes.io/projected/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-kube-api-access-tn6cb\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238514 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238562 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238656 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-combined-ca-bundle\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238720 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238757 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-config\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238802 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr2gj\" (UniqueName: \"kubernetes.io/projected/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-kube-api-access-wr2gj\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.238875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-logs\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.239827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-sb\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.239837 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-swift-storage-0\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.239871 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-svc\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.239995 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-config\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.240555 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-nb\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.309507 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr2gj\" (UniqueName: \"kubernetes.io/projected/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-kube-api-access-wr2gj\") pod \"dnsmasq-dns-6d66f584d7-m48b5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.341184 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn6cb\" (UniqueName: \"kubernetes.io/projected/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-kube-api-access-tn6cb\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.341257 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-combined-ca-bundle\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.341301 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-logs\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.341372 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.341398 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data-custom\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.344727 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-logs\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.347077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-combined-ca-bundle\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.347756 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data-custom\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.349149 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.386595 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn6cb\" (UniqueName: \"kubernetes.io/projected/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-kube-api-access-tn6cb\") pod \"barbican-api-7bc78458f8-mx59q\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.387127 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.617181 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:06 crc kubenswrapper[4744]: I0106 15:00:06.854516 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-74dd54d475-wqvpn"] Jan 06 15:00:07 crc kubenswrapper[4744]: W0106 15:00:07.031014 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ddda95b_00f5_4f93_aa2c_fcf94ea2bc65.slice/crio-5cf8174ae0a171b6d6168b2471644caa3df40634fbcfa94b56fc0fd14dbdc4a9 WatchSource:0}: Error finding container 5cf8174ae0a171b6d6168b2471644caa3df40634fbcfa94b56fc0fd14dbdc4a9: Status 404 returned error can't find the container with id 5cf8174ae0a171b6d6168b2471644caa3df40634fbcfa94b56fc0fd14dbdc4a9 Jan 06 15:00:07 crc kubenswrapper[4744]: I0106 15:00:07.040275 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7bc78458f8-mx59q"] Jan 06 15:00:07 crc kubenswrapper[4744]: I0106 15:00:07.452899 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerStarted","Data":"5cf8174ae0a171b6d6168b2471644caa3df40634fbcfa94b56fc0fd14dbdc4a9"} Jan 06 15:00:07 crc kubenswrapper[4744]: I0106 15:00:07.454058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-74dd54d475-wqvpn" event={"ID":"bc804009-f9fd-4d70-bdd8-c6c61a544549","Type":"ContainerStarted","Data":"ebe87e4ce2da507c33f052a62d16c0898aa0d84aa24c7193540a47aab688d182"} Jan 06 15:00:08 crc kubenswrapper[4744]: I0106 15:00:08.178888 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-m48b5"] Jan 06 15:00:08 crc kubenswrapper[4744]: W0106 15:00:08.275987 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod762cdea3_aba0_4f8f_96c8_4dfaab8bf92a.slice/crio-7a84e4c2a66651c20aa108a6830593619df2ef67ff722d589c3d761c6e49ad09 WatchSource:0}: Error finding container 7a84e4c2a66651c20aa108a6830593619df2ef67ff722d589c3d761c6e49ad09: Status 404 returned error can't find the container with id 7a84e4c2a66651c20aa108a6830593619df2ef67ff722d589c3d761c6e49ad09 Jan 06 15:00:08 crc kubenswrapper[4744]: I0106 15:00:08.277637 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5d48fc6d-cr8dv"] Jan 06 15:00:08 crc kubenswrapper[4744]: I0106 15:00:08.466632 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" event={"ID":"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a","Type":"ContainerStarted","Data":"7a84e4c2a66651c20aa108a6830593619df2ef67ff722d589c3d761c6e49ad09"} Jan 06 15:00:08 crc kubenswrapper[4744]: I0106 15:00:08.468096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" event={"ID":"80cca3b3-144e-4276-80dc-4c0ecb5c34f5","Type":"ContainerStarted","Data":"6f651d90ac7031e2a9f4d3cae611bc86ea45b3a2b1f7a34d9d90f485565246fc"} Jan 06 15:00:08 crc kubenswrapper[4744]: I0106 15:00:08.935275 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6b8c67fcd6-nlf9f"] Jan 06 15:00:08 crc kubenswrapper[4744]: I0106 15:00:08.940433 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:08 crc kubenswrapper[4744]: I0106 15:00:08.942607 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Jan 06 15:00:08 crc kubenswrapper[4744]: I0106 15:00:08.942843 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Jan 06 15:00:08 crc kubenswrapper[4744]: I0106 15:00:08.952105 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b8c67fcd6-nlf9f"] Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.045089 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-internal-tls-certs\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.045151 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-config-data\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.045311 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-combined-ca-bundle\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.045356 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-public-tls-certs\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.045404 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dnb2\" (UniqueName: \"kubernetes.io/projected/c2c0e750-af72-4679-acae-81c93ac6b40e-kube-api-access-6dnb2\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.045566 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2c0e750-af72-4679-acae-81c93ac6b40e-logs\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.045663 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-config-data-custom\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.148319 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-combined-ca-bundle\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.148380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-public-tls-certs\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.148419 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dnb2\" (UniqueName: \"kubernetes.io/projected/c2c0e750-af72-4679-acae-81c93ac6b40e-kube-api-access-6dnb2\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.148454 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2c0e750-af72-4679-acae-81c93ac6b40e-logs\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.148492 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-config-data-custom\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.148666 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-internal-tls-certs\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.148731 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-config-data\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.149300 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2c0e750-af72-4679-acae-81c93ac6b40e-logs\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.155646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-config-data\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.159002 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-internal-tls-certs\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.159131 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-public-tls-certs\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.163996 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-combined-ca-bundle\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.168111 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c2c0e750-af72-4679-acae-81c93ac6b40e-config-data-custom\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.181870 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dnb2\" (UniqueName: \"kubernetes.io/projected/c2c0e750-af72-4679-acae-81c93ac6b40e-kube-api-access-6dnb2\") pod \"barbican-api-6b8c67fcd6-nlf9f\" (UID: \"c2c0e750-af72-4679-acae-81c93ac6b40e\") " pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.260520 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.501480 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerStarted","Data":"200eb467c785f7afde1312ce4bfe4d1dc6b36dc03264f727b5c987e731daf20e"} Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.512613 4744 generic.go:334] "Generic (PLEG): container finished" podID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" containerID="0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa" exitCode=0 Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.512668 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" event={"ID":"80cca3b3-144e-4276-80dc-4c0ecb5c34f5","Type":"ContainerDied","Data":"0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa"} Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.664770 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.772228 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-catalog-content\") pod \"e584b5b6-874e-48ea-b531-267a2c20562b\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.772824 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmfcw\" (UniqueName: \"kubernetes.io/projected/e584b5b6-874e-48ea-b531-267a2c20562b-kube-api-access-hmfcw\") pod \"e584b5b6-874e-48ea-b531-267a2c20562b\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.772933 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-utilities\") pod \"e584b5b6-874e-48ea-b531-267a2c20562b\" (UID: \"e584b5b6-874e-48ea-b531-267a2c20562b\") " Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.774973 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-utilities" (OuterVolumeSpecName: "utilities") pod "e584b5b6-874e-48ea-b531-267a2c20562b" (UID: "e584b5b6-874e-48ea-b531-267a2c20562b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.777547 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e584b5b6-874e-48ea-b531-267a2c20562b-kube-api-access-hmfcw" (OuterVolumeSpecName: "kube-api-access-hmfcw") pod "e584b5b6-874e-48ea-b531-267a2c20562b" (UID: "e584b5b6-874e-48ea-b531-267a2c20562b"). InnerVolumeSpecName "kube-api-access-hmfcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.876011 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmfcw\" (UniqueName: \"kubernetes.io/projected/e584b5b6-874e-48ea-b531-267a2c20562b-kube-api-access-hmfcw\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.876047 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.915671 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e584b5b6-874e-48ea-b531-267a2c20562b" (UID: "e584b5b6-874e-48ea-b531-267a2c20562b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.923241 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b8c67fcd6-nlf9f"] Jan 06 15:00:09 crc kubenswrapper[4744]: I0106 15:00:09.978428 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e584b5b6-874e-48ea-b531-267a2c20562b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.530121 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerStarted","Data":"648a7167f29556e2dc329d67ca06d7e99bc088542f88293040bb18dd4016cb19"} Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.530487 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.532489 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" event={"ID":"c2c0e750-af72-4679-acae-81c93ac6b40e","Type":"ContainerStarted","Data":"4c3066dd728b110a0ea920c0864dfdd754d8bd80591c767293713d23eae53d5d"} Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.535224 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.535262 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.537945 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.548104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pbg7p" event={"ID":"e584b5b6-874e-48ea-b531-267a2c20562b","Type":"ContainerDied","Data":"07b6c3bd77195704a22a462b8b1e61103d0ebf4b80c8cd98eb073df55f197301"} Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.548187 4744 scope.go:117] "RemoveContainer" containerID="f4e9952d12c443bfe8f88057b4ecc3a8985e7e6b4e2e5adae25fff023a27ccaf" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.548270 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pbg7p" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.572509 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.373945974 podStartE2EDuration="31.572490998s" podCreationTimestamp="2026-01-06 14:59:39 +0000 UTC" firstStartedPulling="2026-01-06 14:59:40.005800053 +0000 UTC m=+1376.633266411" lastFinishedPulling="2026-01-06 15:00:09.204345117 +0000 UTC m=+1405.831811435" observedRunningTime="2026-01-06 15:00:10.558377445 +0000 UTC m=+1407.185843763" watchObservedRunningTime="2026-01-06 15:00:10.572490998 +0000 UTC m=+1407.199957316" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.599430 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7bc78458f8-mx59q" podStartSLOduration=4.59940984 podStartE2EDuration="4.59940984s" podCreationTimestamp="2026-01-06 15:00:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:10.583675004 +0000 UTC m=+1407.211141342" watchObservedRunningTime="2026-01-06 15:00:10.59940984 +0000 UTC m=+1407.226876158" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.611804 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" podStartSLOduration=5.611789377 podStartE2EDuration="5.611789377s" podCreationTimestamp="2026-01-06 15:00:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:10.609405914 +0000 UTC m=+1407.236872232" watchObservedRunningTime="2026-01-06 15:00:10.611789377 +0000 UTC m=+1407.239255685" Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.634220 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pbg7p"] Jan 06 15:00:10 crc kubenswrapper[4744]: I0106 15:00:10.643953 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pbg7p"] Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.408103 4744 scope.go:117] "RemoveContainer" containerID="ef9e31b3df62352d36cf9b63fbc7eb73eaba47237a41cce64745243d49776a24" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.418246 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-787d958d8c-44gnt"] Jan 06 15:00:11 crc kubenswrapper[4744]: E0106 15:00:11.418698 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="registry-server" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.418711 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="registry-server" Jan 06 15:00:11 crc kubenswrapper[4744]: E0106 15:00:11.418743 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="extract-utilities" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.418749 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="extract-utilities" Jan 06 15:00:11 crc kubenswrapper[4744]: E0106 15:00:11.418774 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="extract-content" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.418780 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="extract-content" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.418991 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" containerName="registry-server" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.420088 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.422640 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.422857 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.426432 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.432685 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-787d958d8c-44gnt"] Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.467063 4744 scope.go:117] "RemoveContainer" containerID="7b358706df1274187d1b41ce1016e8642b2ae2f7ec73a034eb6f9386e8c12f3f" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.530001 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-public-tls-certs\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.530042 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-config-data\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.530148 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-internal-tls-certs\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.530192 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-combined-ca-bundle\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.530243 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1974504-5e17-4e84-95bd-7d66ed033ccc-etc-swift\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.530268 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1974504-5e17-4e84-95bd-7d66ed033ccc-run-httpd\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.530290 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1974504-5e17-4e84-95bd-7d66ed033ccc-log-httpd\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.530327 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb8nl\" (UniqueName: \"kubernetes.io/projected/e1974504-5e17-4e84-95bd-7d66ed033ccc-kube-api-access-mb8nl\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.562618 4744 generic.go:334] "Generic (PLEG): container finished" podID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerID="9014c95995f56fda3dd0b764d242ac35461935dd633e27558fee29ec99271c66" exitCode=1 Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.562894 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerDied","Data":"9014c95995f56fda3dd0b764d242ac35461935dd633e27558fee29ec99271c66"} Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.563439 4744 scope.go:117] "RemoveContainer" containerID="9014c95995f56fda3dd0b764d242ac35461935dd633e27558fee29ec99271c66" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.565296 4744 generic.go:334] "Generic (PLEG): container finished" podID="5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" containerID="614fe98476c53fb750be24913e7fb2bb300d66d103e2b2ab18c9ac222ce7b9e2" exitCode=0 Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.566402 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6ltjv" event={"ID":"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b","Type":"ContainerDied","Data":"614fe98476c53fb750be24913e7fb2bb300d66d103e2b2ab18c9ac222ce7b9e2"} Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.569972 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" event={"ID":"80cca3b3-144e-4276-80dc-4c0ecb5c34f5","Type":"ContainerStarted","Data":"f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd"} Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.589855 4744 generic.go:334] "Generic (PLEG): container finished" podID="d9e793d1-65f3-43a8-914b-112690afc33f" containerID="70a155e38d038bc8d794f5c4229b8462818b4baae2a1afd8640b477bdf7710e3" exitCode=0 Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.589908 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m88z9" event={"ID":"d9e793d1-65f3-43a8-914b-112690afc33f","Type":"ContainerDied","Data":"70a155e38d038bc8d794f5c4229b8462818b4baae2a1afd8640b477bdf7710e3"} Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.593306 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" event={"ID":"c2c0e750-af72-4679-acae-81c93ac6b40e","Type":"ContainerStarted","Data":"39be715b14f389671f914c864a19aeca714d7d371a7f390526430c9b7f2c077d"} Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.632673 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-internal-tls-certs\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.632769 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-combined-ca-bundle\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.632875 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1974504-5e17-4e84-95bd-7d66ed033ccc-etc-swift\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.632916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1974504-5e17-4e84-95bd-7d66ed033ccc-run-httpd\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.632947 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1974504-5e17-4e84-95bd-7d66ed033ccc-log-httpd\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.633075 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb8nl\" (UniqueName: \"kubernetes.io/projected/e1974504-5e17-4e84-95bd-7d66ed033ccc-kube-api-access-mb8nl\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.633348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-public-tls-certs\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.633380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-config-data\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.634624 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1974504-5e17-4e84-95bd-7d66ed033ccc-run-httpd\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.636453 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1974504-5e17-4e84-95bd-7d66ed033ccc-log-httpd\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.639317 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-internal-tls-certs\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.643909 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e1974504-5e17-4e84-95bd-7d66ed033ccc-etc-swift\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.650413 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-public-tls-certs\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.651042 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-combined-ca-bundle\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.651552 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1974504-5e17-4e84-95bd-7d66ed033ccc-config-data\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.653749 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb8nl\" (UniqueName: \"kubernetes.io/projected/e1974504-5e17-4e84-95bd-7d66ed033ccc-kube-api-access-mb8nl\") pod \"swift-proxy-787d958d8c-44gnt\" (UID: \"e1974504-5e17-4e84-95bd-7d66ed033ccc\") " pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.729086 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e584b5b6-874e-48ea-b531-267a2c20562b" path="/var/lib/kubelet/pods/e584b5b6-874e-48ea-b531-267a2c20562b/volumes" Jan 06 15:00:11 crc kubenswrapper[4744]: I0106 15:00:11.794778 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.389502 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:12 crc kubenswrapper[4744]: W0106 15:00:12.494294 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1974504_5e17_4e84_95bd_7d66ed033ccc.slice/crio-9d267028fcf93846d14469bf6cc4ebf88e1ba5bb65753e6b6da19fac02fe5f92 WatchSource:0}: Error finding container 9d267028fcf93846d14469bf6cc4ebf88e1ba5bb65753e6b6da19fac02fe5f92: Status 404 returned error can't find the container with id 9d267028fcf93846d14469bf6cc4ebf88e1ba5bb65753e6b6da19fac02fe5f92 Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.494877 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-787d958d8c-44gnt"] Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.610592 4744 generic.go:334] "Generic (PLEG): container finished" podID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerID="e9ac5a9c40c477cf86db69b0841546b8eaccb75dbee02e33232bd2cfafdbcae0" exitCode=1 Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.610663 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerDied","Data":"e9ac5a9c40c477cf86db69b0841546b8eaccb75dbee02e33232bd2cfafdbcae0"} Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.610701 4744 scope.go:117] "RemoveContainer" containerID="9014c95995f56fda3dd0b764d242ac35461935dd633e27558fee29ec99271c66" Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.611532 4744 scope.go:117] "RemoveContainer" containerID="e9ac5a9c40c477cf86db69b0841546b8eaccb75dbee02e33232bd2cfafdbcae0" Jan 06 15:00:12 crc kubenswrapper[4744]: E0106 15:00:12.611878 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-7bc78458f8-mx59q_openstack(8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65)\"" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.629845 4744 generic.go:334] "Generic (PLEG): container finished" podID="2b74efd3-9c21-4555-bfa8-5c1652b380dc" containerID="a4deba525591147b2da1e77d8a28191c995aff4a99f4ef3dd3ef23257f19c60a" exitCode=0 Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.629946 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vdnt5" event={"ID":"2b74efd3-9c21-4555-bfa8-5c1652b380dc","Type":"ContainerDied","Data":"a4deba525591147b2da1e77d8a28191c995aff4a99f4ef3dd3ef23257f19c60a"} Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.649844 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" event={"ID":"c2c0e750-af72-4679-acae-81c93ac6b40e","Type":"ContainerStarted","Data":"c23dc3450fe2867991aba926ea2d401a52321801d35c069fba30f95e09fc55ea"} Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.650117 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.650785 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.686007 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" event={"ID":"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a","Type":"ContainerStarted","Data":"9f153192d58e8d31d7bd594c8c4eb0ff6688734a2e5cb6983df219ae91da9781"} Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.690369 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" event={"ID":"762cdea3-aba0-4f8f-96c8-4dfaab8bf92a","Type":"ContainerStarted","Data":"14fe52ca433e019bf45c36a5efcf5384bcc16074d87d2a52927e5fe3e32a050e"} Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.706816 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-74dd54d475-wqvpn" event={"ID":"bc804009-f9fd-4d70-bdd8-c6c61a544549","Type":"ContainerStarted","Data":"b401b5cdc9d13ad55dd93c2b9f8445d40640fe5b7b5fe651289939ebfa9c9947"} Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.706869 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-74dd54d475-wqvpn" event={"ID":"bc804009-f9fd-4d70-bdd8-c6c61a544549","Type":"ContainerStarted","Data":"751c24870f733eb53b8efa43328d491e296947f5278272c8c537793b87d44955"} Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.715944 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-787d958d8c-44gnt" event={"ID":"e1974504-5e17-4e84-95bd-7d66ed033ccc","Type":"ContainerStarted","Data":"9d267028fcf93846d14469bf6cc4ebf88e1ba5bb65753e6b6da19fac02fe5f92"} Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.740195 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" podStartSLOduration=4.740175182 podStartE2EDuration="4.740175182s" podCreationTimestamp="2026-01-06 15:00:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:12.68151444 +0000 UTC m=+1409.308980758" watchObservedRunningTime="2026-01-06 15:00:12.740175182 +0000 UTC m=+1409.367641500" Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.761352 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5d48fc6d-cr8dv" podStartSLOduration=4.544299307 podStartE2EDuration="7.761331741s" podCreationTimestamp="2026-01-06 15:00:05 +0000 UTC" firstStartedPulling="2026-01-06 15:00:08.282602911 +0000 UTC m=+1404.910069229" lastFinishedPulling="2026-01-06 15:00:11.499635345 +0000 UTC m=+1408.127101663" observedRunningTime="2026-01-06 15:00:12.755325432 +0000 UTC m=+1409.382791740" watchObservedRunningTime="2026-01-06 15:00:12.761331741 +0000 UTC m=+1409.388798059" Jan 06 15:00:12 crc kubenswrapper[4744]: I0106 15:00:12.802851 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-74dd54d475-wqvpn" podStartSLOduration=3.145758933 podStartE2EDuration="7.802831519s" podCreationTimestamp="2026-01-06 15:00:05 +0000 UTC" firstStartedPulling="2026-01-06 15:00:06.867426007 +0000 UTC m=+1403.494892325" lastFinishedPulling="2026-01-06 15:00:11.524498593 +0000 UTC m=+1408.151964911" observedRunningTime="2026-01-06 15:00:12.786341413 +0000 UTC m=+1409.413807731" watchObservedRunningTime="2026-01-06 15:00:12.802831519 +0000 UTC m=+1409.430297837" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.236942 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6ltjv" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.291974 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-config-data\") pod \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.292126 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-combined-ca-bundle\") pod \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.292337 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62p9x\" (UniqueName: \"kubernetes.io/projected/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-kube-api-access-62p9x\") pod \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\" (UID: \"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b\") " Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.317732 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-kube-api-access-62p9x" (OuterVolumeSpecName: "kube-api-access-62p9x") pod "5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" (UID: "5ae28705-95b4-4dd8-ab04-bb1e7202ae6b"). InnerVolumeSpecName "kube-api-access-62p9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.353860 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" (UID: "5ae28705-95b4-4dd8-ab04-bb1e7202ae6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.395485 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62p9x\" (UniqueName: \"kubernetes.io/projected/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-kube-api-access-62p9x\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.395518 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.449941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-config-data" (OuterVolumeSpecName: "config-data") pod "5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" (UID: "5ae28705-95b4-4dd8-ab04-bb1e7202ae6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.498617 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.598577 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m88z9" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.705039 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqtcx\" (UniqueName: \"kubernetes.io/projected/d9e793d1-65f3-43a8-914b-112690afc33f-kube-api-access-zqtcx\") pod \"d9e793d1-65f3-43a8-914b-112690afc33f\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.705383 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-combined-ca-bundle\") pod \"d9e793d1-65f3-43a8-914b-112690afc33f\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.705582 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-config\") pod \"d9e793d1-65f3-43a8-914b-112690afc33f\" (UID: \"d9e793d1-65f3-43a8-914b-112690afc33f\") " Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.725301 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9e793d1-65f3-43a8-914b-112690afc33f-kube-api-access-zqtcx" (OuterVolumeSpecName: "kube-api-access-zqtcx") pod "d9e793d1-65f3-43a8-914b-112690afc33f" (UID: "d9e793d1-65f3-43a8-914b-112690afc33f"). InnerVolumeSpecName "kube-api-access-zqtcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.742937 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-m88z9" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.746173 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9e793d1-65f3-43a8-914b-112690afc33f" (UID: "d9e793d1-65f3-43a8-914b-112690afc33f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.748649 4744 scope.go:117] "RemoveContainer" containerID="e9ac5a9c40c477cf86db69b0841546b8eaccb75dbee02e33232bd2cfafdbcae0" Jan 06 15:00:13 crc kubenswrapper[4744]: E0106 15:00:13.748833 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-7bc78458f8-mx59q_openstack(8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65)\"" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.750707 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6ltjv" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.791849 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-787d958d8c-44gnt" podStartSLOduration=2.791828153 podStartE2EDuration="2.791828153s" podCreationTimestamp="2026-01-06 15:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:13.781640934 +0000 UTC m=+1410.409107252" watchObservedRunningTime="2026-01-06 15:00:13.791828153 +0000 UTC m=+1410.419294461" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.793326 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-config" (OuterVolumeSpecName: "config") pod "d9e793d1-65f3-43a8-914b-112690afc33f" (UID: "d9e793d1-65f3-43a8-914b-112690afc33f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.810916 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.810961 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqtcx\" (UniqueName: \"kubernetes.io/projected/d9e793d1-65f3-43a8-914b-112690afc33f-kube-api-access-zqtcx\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.810976 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9e793d1-65f3-43a8-914b-112690afc33f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.888982 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-m88z9" event={"ID":"d9e793d1-65f3-43a8-914b-112690afc33f","Type":"ContainerDied","Data":"e1ae18628c2b1d69913b1b9e490361f02027a74b017bdbd6e8bdd7119016a621"} Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.889024 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1ae18628c2b1d69913b1b9e490361f02027a74b017bdbd6e8bdd7119016a621" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.889043 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-787d958d8c-44gnt" event={"ID":"e1974504-5e17-4e84-95bd-7d66ed033ccc","Type":"ContainerStarted","Data":"7ef9c5875b2f8b1551fe5e89ccfc91513bdb87a3d7ef745d96bf9dff309a7d7a"} Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.889057 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-787d958d8c-44gnt" event={"ID":"e1974504-5e17-4e84-95bd-7d66ed033ccc","Type":"ContainerStarted","Data":"aed593cd774dbeeeb656d86ef64776debf32efeaa51c18c828d15100f06064ab"} Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.889075 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.889095 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.889104 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6ltjv" event={"ID":"5ae28705-95b4-4dd8-ab04-bb1e7202ae6b","Type":"ContainerDied","Data":"99bcb3b0f6adef94a66559e37041aec8a53c8c8dcf357f45ed22d76e6f7be89f"} Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.889116 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99bcb3b0f6adef94a66559e37041aec8a53c8c8dcf357f45ed22d76e6f7be89f" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.911448 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-m48b5"] Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.911668 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" podUID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" containerName="dnsmasq-dns" containerID="cri-o://f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd" gracePeriod=10 Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.937078 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dc6576bbb-qn5wg"] Jan 06 15:00:13 crc kubenswrapper[4744]: E0106 15:00:13.937528 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9e793d1-65f3-43a8-914b-112690afc33f" containerName="neutron-db-sync" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.937545 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9e793d1-65f3-43a8-914b-112690afc33f" containerName="neutron-db-sync" Jan 06 15:00:13 crc kubenswrapper[4744]: E0106 15:00:13.937558 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" containerName="heat-db-sync" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.937565 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" containerName="heat-db-sync" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.937763 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9e793d1-65f3-43a8-914b-112690afc33f" containerName="neutron-db-sync" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.937782 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" containerName="heat-db-sync" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.938860 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.944078 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.954212 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dc6576bbb-qn5wg"] Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.965527 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-m9cb7"] Jan 06 15:00:13 crc kubenswrapper[4744]: I0106 15:00:13.969351 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.014525 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.014871 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-combined-ca-bundle\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.014904 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8dqt\" (UniqueName: \"kubernetes.io/projected/1c2a7970-7262-45b0-91c4-96b457f06692-kube-api-access-s8dqt\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.014926 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-config\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.014950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.014989 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4jgd\" (UniqueName: \"kubernetes.io/projected/826e0141-c6b9-46bc-8450-88cef322d974-kube-api-access-m4jgd\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.015034 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-ovndb-tls-certs\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.015083 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.015118 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-svc\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.015197 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-config\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.015215 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-httpd-config\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.022879 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-m9cb7"] Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.124776 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-config\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.124837 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-httpd-config\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.124880 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.124924 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-combined-ca-bundle\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.124951 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8dqt\" (UniqueName: \"kubernetes.io/projected/1c2a7970-7262-45b0-91c4-96b457f06692-kube-api-access-s8dqt\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.124972 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-config\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.124997 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.125038 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4jgd\" (UniqueName: \"kubernetes.io/projected/826e0141-c6b9-46bc-8450-88cef322d974-kube-api-access-m4jgd\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.125085 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-ovndb-tls-certs\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.125133 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.125248 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-svc\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.125943 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.125954 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-svc\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.126752 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.127028 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-config\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.127314 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.133503 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.135039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-combined-ca-bundle\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.140059 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-ovndb-tls-certs\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.143737 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-httpd-config\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.145948 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-config\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.160845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8dqt\" (UniqueName: \"kubernetes.io/projected/1c2a7970-7262-45b0-91c4-96b457f06692-kube-api-access-s8dqt\") pod \"dnsmasq-dns-688c87cc99-m9cb7\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.161036 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4jgd\" (UniqueName: \"kubernetes.io/projected/826e0141-c6b9-46bc-8450-88cef322d974-kube-api-access-m4jgd\") pod \"neutron-dc6576bbb-qn5wg\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.271551 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.300694 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.727248 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vdnt5" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.754587 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.801605 4744 generic.go:334] "Generic (PLEG): container finished" podID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" containerID="f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd" exitCode=0 Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.801679 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" event={"ID":"80cca3b3-144e-4276-80dc-4c0ecb5c34f5","Type":"ContainerDied","Data":"f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd"} Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.801704 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" event={"ID":"80cca3b3-144e-4276-80dc-4c0ecb5c34f5","Type":"ContainerDied","Data":"6f651d90ac7031e2a9f4d3cae611bc86ea45b3a2b1f7a34d9d90f485565246fc"} Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.801723 4744 scope.go:117] "RemoveContainer" containerID="f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.801883 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.833338 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-vdnt5" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.833734 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-vdnt5" event={"ID":"2b74efd3-9c21-4555-bfa8-5c1652b380dc","Type":"ContainerDied","Data":"874175b57595591419a94ec9fceaadcc097a204bb4ce91d8ac4a450c636a87dc"} Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.833799 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="874175b57595591419a94ec9fceaadcc097a204bb4ce91d8ac4a450c636a87dc" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.859347 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbzxc\" (UniqueName: \"kubernetes.io/projected/2b74efd3-9c21-4555-bfa8-5c1652b380dc-kube-api-access-bbzxc\") pod \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.859430 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-combined-ca-bundle\") pod \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.859481 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-scripts\") pod \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.859538 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr2gj\" (UniqueName: \"kubernetes.io/projected/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-kube-api-access-wr2gj\") pod \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.859558 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-sb\") pod \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.859678 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-db-sync-config-data\") pod \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.867381 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-config-data\") pod \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.867975 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-nb\") pod \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.868308 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-swift-storage-0\") pod \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.868412 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-svc\") pod \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.868562 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b74efd3-9c21-4555-bfa8-5c1652b380dc-etc-machine-id\") pod \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\" (UID: \"2b74efd3-9c21-4555-bfa8-5c1652b380dc\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.868592 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-config\") pod \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\" (UID: \"80cca3b3-144e-4276-80dc-4c0ecb5c34f5\") " Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.903079 4744 scope.go:117] "RemoveContainer" containerID="0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.905182 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b74efd3-9c21-4555-bfa8-5c1652b380dc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2b74efd3-9c21-4555-bfa8-5c1652b380dc" (UID: "2b74efd3-9c21-4555-bfa8-5c1652b380dc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 15:00:14 crc kubenswrapper[4744]: I0106 15:00:14.985683 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b74efd3-9c21-4555-bfa8-5c1652b380dc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.105468 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-m9cb7"] Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.384253 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-kube-api-access-wr2gj" (OuterVolumeSpecName: "kube-api-access-wr2gj") pod "80cca3b3-144e-4276-80dc-4c0ecb5c34f5" (UID: "80cca3b3-144e-4276-80dc-4c0ecb5c34f5"). InnerVolumeSpecName "kube-api-access-wr2gj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.384410 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-scripts" (OuterVolumeSpecName: "scripts") pod "2b74efd3-9c21-4555-bfa8-5c1652b380dc" (UID: "2b74efd3-9c21-4555-bfa8-5c1652b380dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.385454 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b74efd3-9c21-4555-bfa8-5c1652b380dc-kube-api-access-bbzxc" (OuterVolumeSpecName: "kube-api-access-bbzxc") pod "2b74efd3-9c21-4555-bfa8-5c1652b380dc" (UID: "2b74efd3-9c21-4555-bfa8-5c1652b380dc"). InnerVolumeSpecName "kube-api-access-bbzxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.386764 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2b74efd3-9c21-4555-bfa8-5c1652b380dc" (UID: "2b74efd3-9c21-4555-bfa8-5c1652b380dc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.388867 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.389845 4744 scope.go:117] "RemoveContainer" containerID="e9ac5a9c40c477cf86db69b0841546b8eaccb75dbee02e33232bd2cfafdbcae0" Jan 06 15:00:15 crc kubenswrapper[4744]: E0106 15:00:15.390227 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-7bc78458f8-mx59q_openstack(8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65)\"" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.393243 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.393347 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.395341 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbzxc\" (UniqueName: \"kubernetes.io/projected/2b74efd3-9c21-4555-bfa8-5c1652b380dc-kube-api-access-bbzxc\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.395367 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.395380 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr2gj\" (UniqueName: \"kubernetes.io/projected/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-kube-api-access-wr2gj\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.395391 4744 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.413070 4744 scope.go:117] "RemoveContainer" containerID="f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd" Jan 06 15:00:15 crc kubenswrapper[4744]: E0106 15:00:15.424739 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd\": container with ID starting with f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd not found: ID does not exist" containerID="f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.424779 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd"} err="failed to get container status \"f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd\": rpc error: code = NotFound desc = could not find container \"f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd\": container with ID starting with f0d897c46eb904380974c10d3cf4a7fa045421a4181e746251078f8d0d1458fd not found: ID does not exist" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.424805 4744 scope.go:117] "RemoveContainer" containerID="0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa" Jan 06 15:00:15 crc kubenswrapper[4744]: E0106 15:00:15.431278 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa\": container with ID starting with 0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa not found: ID does not exist" containerID="0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.431328 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa"} err="failed to get container status \"0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa\": rpc error: code = NotFound desc = could not find container \"0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa\": container with ID starting with 0fb1a823f06e9871fd536cd3fa5a7c1ca1dfb0d2608924c120fcae5e070716aa not found: ID does not exist" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.437995 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.438291 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="ceilometer-central-agent" containerID="cri-o://c9e09149319fbba6a28036746c6d5bddf814c68e2b770403b877daccfcd20920" gracePeriod=30 Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.438599 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="sg-core" containerID="cri-o://bdd65b64eac259914eb01ea6b6212e354d572e83d74fb6723c6b5db93e2ead7f" gracePeriod=30 Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.438730 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="proxy-httpd" containerID="cri-o://648a7167f29556e2dc329d67ca06d7e99bc088542f88293040bb18dd4016cb19" gracePeriod=30 Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.438780 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="ceilometer-notification-agent" containerID="cri-o://fe80bcca808b45ed0e1be2e4996875a0a96b62f5701c5854325766fa27f720af" gracePeriod=30 Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.487244 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "80cca3b3-144e-4276-80dc-4c0ecb5c34f5" (UID: "80cca3b3-144e-4276-80dc-4c0ecb5c34f5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.487310 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b74efd3-9c21-4555-bfa8-5c1652b380dc" (UID: "2b74efd3-9c21-4555-bfa8-5c1652b380dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.487988 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-config" (OuterVolumeSpecName: "config") pod "80cca3b3-144e-4276-80dc-4c0ecb5c34f5" (UID: "80cca3b3-144e-4276-80dc-4c0ecb5c34f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.510089 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.510747 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.510768 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.511581 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "80cca3b3-144e-4276-80dc-4c0ecb5c34f5" (UID: "80cca3b3-144e-4276-80dc-4c0ecb5c34f5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.522730 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "80cca3b3-144e-4276-80dc-4c0ecb5c34f5" (UID: "80cca3b3-144e-4276-80dc-4c0ecb5c34f5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.603901 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-config-data" (OuterVolumeSpecName: "config-data") pod "2b74efd3-9c21-4555-bfa8-5c1652b380dc" (UID: "2b74efd3-9c21-4555-bfa8-5c1652b380dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.613383 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b74efd3-9c21-4555-bfa8-5c1652b380dc-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.613411 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.613422 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.681940 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "80cca3b3-144e-4276-80dc-4c0ecb5c34f5" (UID: "80cca3b3-144e-4276-80dc-4c0ecb5c34f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.682629 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dc6576bbb-qn5wg"] Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.714879 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/80cca3b3-144e-4276-80dc-4c0ecb5c34f5-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.862401 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc6576bbb-qn5wg" event={"ID":"826e0141-c6b9-46bc-8450-88cef322d974","Type":"ContainerStarted","Data":"2b06a91900b437fe3ccafd75d3bf8483ade95942fcff3763072c6fbcdd086cf4"} Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.884752 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" event={"ID":"1c2a7970-7262-45b0-91c4-96b457f06692","Type":"ContainerStarted","Data":"4adde1f274bd4d4b703f1d3c79b58f0d8400b7f652335914146592c43df3bb67"} Jan 06 15:00:15 crc kubenswrapper[4744]: E0106 15:00:15.975605 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7966af7_509d_4b51_aa77_35a21616df45.slice/crio-648a7167f29556e2dc329d67ca06d7e99bc088542f88293040bb18dd4016cb19.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7966af7_509d_4b51_aa77_35a21616df45.slice/crio-bdd65b64eac259914eb01ea6b6212e354d572e83d74fb6723c6b5db93e2ead7f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80cca3b3_144e_4276_80dc_4c0ecb5c34f5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b74efd3_9c21_4555_bfa8_5c1652b380dc.slice/crio-874175b57595591419a94ec9fceaadcc097a204bb4ce91d8ac4a450c636a87dc\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b74efd3_9c21_4555_bfa8_5c1652b380dc.slice\": RecentStats: unable to find data in memory cache]" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.992229 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 06 15:00:15 crc kubenswrapper[4744]: E0106 15:00:15.992877 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" containerName="dnsmasq-dns" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.992900 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" containerName="dnsmasq-dns" Jan 06 15:00:15 crc kubenswrapper[4744]: E0106 15:00:15.992934 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b74efd3-9c21-4555-bfa8-5c1652b380dc" containerName="cinder-db-sync" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.992944 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b74efd3-9c21-4555-bfa8-5c1652b380dc" containerName="cinder-db-sync" Jan 06 15:00:15 crc kubenswrapper[4744]: E0106 15:00:15.992962 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" containerName="init" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.992970 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" containerName="init" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.993291 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b74efd3-9c21-4555-bfa8-5c1652b380dc" containerName="cinder-db-sync" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.993315 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" containerName="dnsmasq-dns" Jan 06 15:00:15 crc kubenswrapper[4744]: I0106 15:00:15.994847 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.010994 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.011241 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.013852 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.014085 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-7hzlw" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.038304 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.131519 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.131843 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1b118af-7005-4a34-94ae-cf00f8c63839-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.131863 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px7nn\" (UniqueName: \"kubernetes.io/projected/a1b118af-7005-4a34-94ae-cf00f8c63839-kube-api-access-px7nn\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.131895 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-scripts\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.131939 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.131964 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.136353 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-m9cb7"] Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.149254 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-znjwg"] Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.151061 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.160469 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-znjwg"] Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.232105 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.234744 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.234806 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.234832 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.234857 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.234894 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1b118af-7005-4a34-94ae-cf00f8c63839-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.234916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px7nn\" (UniqueName: \"kubernetes.io/projected/a1b118af-7005-4a34-94ae-cf00f8c63839-kube-api-access-px7nn\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.234939 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5cl7\" (UniqueName: \"kubernetes.io/projected/bf6d608b-355b-471e-9fab-40b62a25b8f0-kube-api-access-k5cl7\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.234958 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.234986 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-scripts\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.235032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.235057 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.235179 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.235231 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-config\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.235919 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1b118af-7005-4a34-94ae-cf00f8c63839-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.242444 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.243632 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.249027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.249425 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.249883 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.251208 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-scripts\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.262675 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px7nn\" (UniqueName: \"kubernetes.io/projected/a1b118af-7005-4a34-94ae-cf00f8c63839-kube-api-access-px7nn\") pod \"cinder-scheduler-0\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.337048 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-scripts\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.337111 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqdrl\" (UniqueName: \"kubernetes.io/projected/2c546315-1cd0-45e3-809b-c6a8d2e509b3-kube-api-access-pqdrl\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.337147 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.337204 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.337727 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.338088 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.338169 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-config\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.338205 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.338222 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.338244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.338306 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data-custom\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.338370 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5cl7\" (UniqueName: \"kubernetes.io/projected/bf6d608b-355b-471e-9fab-40b62a25b8f0-kube-api-access-k5cl7\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.338399 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.338426 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c546315-1cd0-45e3-809b-c6a8d2e509b3-logs\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.339004 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.339027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.339251 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.339405 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c546315-1cd0-45e3-809b-c6a8d2e509b3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.339413 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-config\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.354011 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5cl7\" (UniqueName: \"kubernetes.io/projected/bf6d608b-355b-471e-9fab-40b62a25b8f0-kube-api-access-k5cl7\") pod \"dnsmasq-dns-6bb4fc677f-znjwg\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.395948 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.398080 4744 scope.go:117] "RemoveContainer" containerID="e9ac5a9c40c477cf86db69b0841546b8eaccb75dbee02e33232bd2cfafdbcae0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.398296 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.398658 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:16 crc kubenswrapper[4744]: E0106 15:00:16.399124 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=barbican-api pod=barbican-api-7bc78458f8-mx59q_openstack(8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65)\"" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.400961 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c5f6db8dc-vwrjb"] Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.432826 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c5f6db8dc-vwrjb"] Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.435716 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.442909 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.442987 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.449686 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data-custom\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.450368 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c546315-1cd0-45e3-809b-c6a8d2e509b3-logs\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.450569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c546315-1cd0-45e3-809b-c6a8d2e509b3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.450742 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.450778 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-scripts\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.450839 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqdrl\" (UniqueName: \"kubernetes.io/projected/2c546315-1cd0-45e3-809b-c6a8d2e509b3-kube-api-access-pqdrl\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.450916 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.451286 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c546315-1cd0-45e3-809b-c6a8d2e509b3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.451733 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c546315-1cd0-45e3-809b-c6a8d2e509b3-logs\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.453542 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.453735 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data-custom\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.462200 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-scripts\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.465763 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.478208 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqdrl\" (UniqueName: \"kubernetes.io/projected/2c546315-1cd0-45e3-809b-c6a8d2e509b3-kube-api-access-pqdrl\") pod \"cinder-api-0\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.478897 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.553104 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-ovndb-tls-certs\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.553445 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-internal-tls-certs\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.553484 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-combined-ca-bundle\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.553550 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr26k\" (UniqueName: \"kubernetes.io/projected/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-kube-api-access-fr26k\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.553574 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-public-tls-certs\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.553595 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-httpd-config\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.553695 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-config\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.586273 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.662112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-config\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.662495 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-ovndb-tls-certs\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.662524 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-internal-tls-certs\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.662563 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-combined-ca-bundle\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.663262 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr26k\" (UniqueName: \"kubernetes.io/projected/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-kube-api-access-fr26k\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.663304 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-public-tls-certs\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.663334 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-httpd-config\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.684296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-config\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.685097 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-internal-tls-certs\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.685579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-ovndb-tls-certs\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.694136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-httpd-config\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.699075 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-public-tls-certs\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.703980 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-combined-ca-bundle\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.712325 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr26k\" (UniqueName: \"kubernetes.io/projected/ab974dd7-0476-4bec-8c0d-b7bdcf52470f-kube-api-access-fr26k\") pod \"neutron-5c5f6db8dc-vwrjb\" (UID: \"ab974dd7-0476-4bec-8c0d-b7bdcf52470f\") " pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.903258 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.923994 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc6576bbb-qn5wg" event={"ID":"826e0141-c6b9-46bc-8450-88cef322d974","Type":"ContainerStarted","Data":"4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3"} Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.942583 4744 generic.go:334] "Generic (PLEG): container finished" podID="1c2a7970-7262-45b0-91c4-96b457f06692" containerID="eba3c052c41df330e8d6cad2e89b12f695191aa8f452bf04dfb492c471925eca" exitCode=0 Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.942987 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" event={"ID":"1c2a7970-7262-45b0-91c4-96b457f06692","Type":"ContainerDied","Data":"eba3c052c41df330e8d6cad2e89b12f695191aa8f452bf04dfb492c471925eca"} Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.984169 4744 generic.go:334] "Generic (PLEG): container finished" podID="c7966af7-509d-4b51-aa77-35a21616df45" containerID="648a7167f29556e2dc329d67ca06d7e99bc088542f88293040bb18dd4016cb19" exitCode=0 Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.984199 4744 generic.go:334] "Generic (PLEG): container finished" podID="c7966af7-509d-4b51-aa77-35a21616df45" containerID="bdd65b64eac259914eb01ea6b6212e354d572e83d74fb6723c6b5db93e2ead7f" exitCode=2 Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.984205 4744 generic.go:334] "Generic (PLEG): container finished" podID="c7966af7-509d-4b51-aa77-35a21616df45" containerID="fe80bcca808b45ed0e1be2e4996875a0a96b62f5701c5854325766fa27f720af" exitCode=0 Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.984214 4744 generic.go:334] "Generic (PLEG): container finished" podID="c7966af7-509d-4b51-aa77-35a21616df45" containerID="c9e09149319fbba6a28036746c6d5bddf814c68e2b770403b877daccfcd20920" exitCode=0 Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.984235 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerDied","Data":"648a7167f29556e2dc329d67ca06d7e99bc088542f88293040bb18dd4016cb19"} Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.984260 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerDied","Data":"bdd65b64eac259914eb01ea6b6212e354d572e83d74fb6723c6b5db93e2ead7f"} Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.984277 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerDied","Data":"fe80bcca808b45ed0e1be2e4996875a0a96b62f5701c5854325766fa27f720af"} Jan 06 15:00:16 crc kubenswrapper[4744]: I0106 15:00:16.984287 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerDied","Data":"c9e09149319fbba6a28036746c6d5bddf814c68e2b770403b877daccfcd20920"} Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.009337 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.166792 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.167386 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerName="glance-log" containerID="cri-o://0d55873db7667c2a58e81e8697f3b88667d3f505075d94fb72f3e82451ef5938" gracePeriod=30 Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.167847 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerName="glance-httpd" containerID="cri-o://9281ea95ced4d12abbc3f13ffac89b4f12c76d9e8363006875383c1c108ad996" gracePeriod=30 Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.504508 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 06 15:00:17 crc kubenswrapper[4744]: W0106 15:00:17.524043 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-7a1eabb1fd46fa9eaa70c0863168b519b6bb07624f9e45e231fc57023afbe68d WatchSource:0}: Error finding container 7a1eabb1fd46fa9eaa70c0863168b519b6bb07624f9e45e231fc57023afbe68d: Status 404 returned error can't find the container with id 7a1eabb1fd46fa9eaa70c0863168b519b6bb07624f9e45e231fc57023afbe68d Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.540894 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-znjwg"] Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.789201 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.924869 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-log-httpd\") pod \"c7966af7-509d-4b51-aa77-35a21616df45\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.924952 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-run-httpd\") pod \"c7966af7-509d-4b51-aa77-35a21616df45\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.925178 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-config-data\") pod \"c7966af7-509d-4b51-aa77-35a21616df45\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.925205 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-combined-ca-bundle\") pod \"c7966af7-509d-4b51-aa77-35a21616df45\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.925225 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhfkr\" (UniqueName: \"kubernetes.io/projected/c7966af7-509d-4b51-aa77-35a21616df45-kube-api-access-jhfkr\") pod \"c7966af7-509d-4b51-aa77-35a21616df45\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.925264 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-sg-core-conf-yaml\") pod \"c7966af7-509d-4b51-aa77-35a21616df45\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.925293 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-scripts\") pod \"c7966af7-509d-4b51-aa77-35a21616df45\" (UID: \"c7966af7-509d-4b51-aa77-35a21616df45\") " Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.925560 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c7966af7-509d-4b51-aa77-35a21616df45" (UID: "c7966af7-509d-4b51-aa77-35a21616df45"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.925796 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.927471 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c7966af7-509d-4b51-aa77-35a21616df45" (UID: "c7966af7-509d-4b51-aa77-35a21616df45"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.938389 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-scripts" (OuterVolumeSpecName: "scripts") pod "c7966af7-509d-4b51-aa77-35a21616df45" (UID: "c7966af7-509d-4b51-aa77-35a21616df45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.958519 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7966af7-509d-4b51-aa77-35a21616df45-kube-api-access-jhfkr" (OuterVolumeSpecName: "kube-api-access-jhfkr") pod "c7966af7-509d-4b51-aa77-35a21616df45" (UID: "c7966af7-509d-4b51-aa77-35a21616df45"). InnerVolumeSpecName "kube-api-access-jhfkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:17 crc kubenswrapper[4744]: I0106 15:00:17.984452 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c5f6db8dc-vwrjb"] Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:17.998212 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c7966af7-509d-4b51-aa77-35a21616df45" (UID: "c7966af7-509d-4b51-aa77-35a21616df45"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.029344 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7966af7-509d-4b51-aa77-35a21616df45-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.029374 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhfkr\" (UniqueName: \"kubernetes.io/projected/c7966af7-509d-4b51-aa77-35a21616df45-kube-api-access-jhfkr\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.029387 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.029400 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.094904 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" event={"ID":"bf6d608b-355b-471e-9fab-40b62a25b8f0","Type":"ContainerStarted","Data":"6fb418c65bf1ed1af72f36bd8c4cb20d78214bcc63a2c286c95b3c5da6c36dc5"} Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.096730 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a1b118af-7005-4a34-94ae-cf00f8c63839","Type":"ContainerStarted","Data":"0536095e73f584c3907f1a4fa7e17fc39425f2fba801b6f7da26ea8a35c6bd52"} Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.106229 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7966af7-509d-4b51-aa77-35a21616df45","Type":"ContainerDied","Data":"99fad64f1b9c5778596c0d95f19d13f49f060be6f388c723a1e718cb4c329f4c"} Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.106280 4744 scope.go:117] "RemoveContainer" containerID="648a7167f29556e2dc329d67ca06d7e99bc088542f88293040bb18dd4016cb19" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.106407 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.121344 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2c546315-1cd0-45e3-809b-c6a8d2e509b3","Type":"ContainerStarted","Data":"7a1eabb1fd46fa9eaa70c0863168b519b6bb07624f9e45e231fc57023afbe68d"} Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.136105 4744 generic.go:334] "Generic (PLEG): container finished" podID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerID="0d55873db7667c2a58e81e8697f3b88667d3f505075d94fb72f3e82451ef5938" exitCode=143 Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.136190 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cab7dd39-8d15-462c-9df0-dc4df7e54db6","Type":"ContainerDied","Data":"0d55873db7667c2a58e81e8697f3b88667d3f505075d94fb72f3e82451ef5938"} Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.142091 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc6576bbb-qn5wg" event={"ID":"826e0141-c6b9-46bc-8450-88cef322d974","Type":"ContainerStarted","Data":"311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489"} Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.142132 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.146363 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-config-data" (OuterVolumeSpecName: "config-data") pod "c7966af7-509d-4b51-aa77-35a21616df45" (UID: "c7966af7-509d-4b51-aa77-35a21616df45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.167388 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7966af7-509d-4b51-aa77-35a21616df45" (UID: "c7966af7-509d-4b51-aa77-35a21616df45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.182011 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dc6576bbb-qn5wg" podStartSLOduration=5.181986391 podStartE2EDuration="5.181986391s" podCreationTimestamp="2026-01-06 15:00:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:18.165065103 +0000 UTC m=+1414.792531421" watchObservedRunningTime="2026-01-06 15:00:18.181986391 +0000 UTC m=+1414.809452709" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.239561 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.239591 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7966af7-509d-4b51-aa77-35a21616df45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.321652 4744 scope.go:117] "RemoveContainer" containerID="bdd65b64eac259914eb01ea6b6212e354d572e83d74fb6723c6b5db93e2ead7f" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.389475 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.436208 4744 scope.go:117] "RemoveContainer" containerID="fe80bcca808b45ed0e1be2e4996875a0a96b62f5701c5854325766fa27f720af" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.647719 4744 scope.go:117] "RemoveContainer" containerID="c9e09149319fbba6a28036746c6d5bddf814c68e2b770403b877daccfcd20920" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.757200 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.766356 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.785686 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:18 crc kubenswrapper[4744]: E0106 15:00:18.786139 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="ceilometer-notification-agent" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.786172 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="ceilometer-notification-agent" Jan 06 15:00:18 crc kubenswrapper[4744]: E0106 15:00:18.786204 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="sg-core" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.786210 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="sg-core" Jan 06 15:00:18 crc kubenswrapper[4744]: E0106 15:00:18.786220 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="proxy-httpd" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.786227 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="proxy-httpd" Jan 06 15:00:18 crc kubenswrapper[4744]: E0106 15:00:18.786237 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="ceilometer-central-agent" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.786243 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="ceilometer-central-agent" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.786430 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="sg-core" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.786455 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="ceilometer-notification-agent" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.786463 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="proxy-httpd" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.786480 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7966af7-509d-4b51-aa77-35a21616df45" containerName="ceilometer-central-agent" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.788508 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.790074 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.822818 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.841141 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.866586 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-log-httpd\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.866659 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.866738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p96t\" (UniqueName: \"kubernetes.io/projected/c0405b86-e930-436b-94a0-14b08f60f6c1-kube-api-access-2p96t\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.866760 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.866935 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-config-data\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.866980 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-run-httpd\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.867061 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-scripts\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.882061 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.968632 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-svc\") pod \"1c2a7970-7262-45b0-91c4-96b457f06692\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.968700 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-nb\") pod \"1c2a7970-7262-45b0-91c4-96b457f06692\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.968914 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8dqt\" (UniqueName: \"kubernetes.io/projected/1c2a7970-7262-45b0-91c4-96b457f06692-kube-api-access-s8dqt\") pod \"1c2a7970-7262-45b0-91c4-96b457f06692\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.968975 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-sb\") pod \"1c2a7970-7262-45b0-91c4-96b457f06692\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.969031 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-swift-storage-0\") pod \"1c2a7970-7262-45b0-91c4-96b457f06692\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.969135 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-config\") pod \"1c2a7970-7262-45b0-91c4-96b457f06692\" (UID: \"1c2a7970-7262-45b0-91c4-96b457f06692\") " Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.969488 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-scripts\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.969545 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-log-httpd\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.969594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.969663 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p96t\" (UniqueName: \"kubernetes.io/projected/c0405b86-e930-436b-94a0-14b08f60f6c1-kube-api-access-2p96t\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.969685 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.972008 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-log-httpd\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.976800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.985883 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-config-data\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.985931 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-run-httpd\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.986399 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-run-httpd\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.996897 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:18 crc kubenswrapper[4744]: I0106 15:00:18.999218 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-config-data\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.001018 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-scripts\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.012936 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c2a7970-7262-45b0-91c4-96b457f06692-kube-api-access-s8dqt" (OuterVolumeSpecName: "kube-api-access-s8dqt") pod "1c2a7970-7262-45b0-91c4-96b457f06692" (UID: "1c2a7970-7262-45b0-91c4-96b457f06692"). InnerVolumeSpecName "kube-api-access-s8dqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.021169 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p96t\" (UniqueName: \"kubernetes.io/projected/c0405b86-e930-436b-94a0-14b08f60f6c1-kube-api-access-2p96t\") pod \"ceilometer-0\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " pod="openstack/ceilometer-0" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.023897 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1c2a7970-7262-45b0-91c4-96b457f06692" (UID: "1c2a7970-7262-45b0-91c4-96b457f06692"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.027846 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1c2a7970-7262-45b0-91c4-96b457f06692" (UID: "1c2a7970-7262-45b0-91c4-96b457f06692"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.036850 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1c2a7970-7262-45b0-91c4-96b457f06692" (UID: "1c2a7970-7262-45b0-91c4-96b457f06692"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.039697 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1c2a7970-7262-45b0-91c4-96b457f06692" (UID: "1c2a7970-7262-45b0-91c4-96b457f06692"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.061782 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-config" (OuterVolumeSpecName: "config") pod "1c2a7970-7262-45b0-91c4-96b457f06692" (UID: "1c2a7970-7262-45b0-91c4-96b457f06692"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.088060 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.088091 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.088101 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.088115 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8dqt\" (UniqueName: \"kubernetes.io/projected/1c2a7970-7262-45b0-91c4-96b457f06692-kube-api-access-s8dqt\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.088123 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.088131 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1c2a7970-7262-45b0-91c4-96b457f06692-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.172029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c5f6db8dc-vwrjb" event={"ID":"ab974dd7-0476-4bec-8c0d-b7bdcf52470f","Type":"ContainerStarted","Data":"6fd11e618010dd1221d9b2728e85d6626d06eeadd658c73f06a6ec630207dff7"} Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.175605 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf6d608b-355b-471e-9fab-40b62a25b8f0" containerID="845cc2c773efac6f0e9deb2039d627164a3490782ca8e57484b4b28d025efc01" exitCode=0 Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.175646 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" event={"ID":"bf6d608b-355b-471e-9fab-40b62a25b8f0","Type":"ContainerDied","Data":"845cc2c773efac6f0e9deb2039d627164a3490782ca8e57484b4b28d025efc01"} Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.187864 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" event={"ID":"1c2a7970-7262-45b0-91c4-96b457f06692","Type":"ContainerDied","Data":"4adde1f274bd4d4b703f1d3c79b58f0d8400b7f652335914146592c43df3bb67"} Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.187923 4744 scope.go:117] "RemoveContainer" containerID="eba3c052c41df330e8d6cad2e89b12f695191aa8f452bf04dfb492c471925eca" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.187884 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-m9cb7" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.206394 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.339349 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-m9cb7"] Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.373564 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-m9cb7"] Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.610669 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.611095 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="992860bb-2626-401b-b4a0-439bb834f8e3" containerName="glance-log" containerID="cri-o://f704408826be9ed92a9d2c0dfade364364a46376f5afb28dfca7b651afcd64f2" gracePeriod=30 Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.611379 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="992860bb-2626-401b-b4a0-439bb834f8e3" containerName="glance-httpd" containerID="cri-o://de4be668780cd37bacbeb7ad79259bca72d6ec15130dcb1a90d59bd1ee307017" gracePeriod=30 Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.768871 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c2a7970-7262-45b0-91c4-96b457f06692" path="/var/lib/kubelet/pods/1c2a7970-7262-45b0-91c4-96b457f06692/volumes" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.769859 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7966af7-509d-4b51-aa77-35a21616df45" path="/var/lib/kubelet/pods/c7966af7-509d-4b51-aa77-35a21616df45/volumes" Jan 06 15:00:19 crc kubenswrapper[4744]: I0106 15:00:19.966226 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.000796 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.212766 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c5f6db8dc-vwrjb" event={"ID":"ab974dd7-0476-4bec-8c0d-b7bdcf52470f","Type":"ContainerStarted","Data":"84d89eb298e6fb5925039deac284cb8510c6d439c86839deae794fe3ef8fbf91"} Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.213095 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c5f6db8dc-vwrjb" event={"ID":"ab974dd7-0476-4bec-8c0d-b7bdcf52470f","Type":"ContainerStarted","Data":"b60c401c64d8f43cf206c2374e673b12c25e97e96f44d781eddca5d721b226e5"} Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.218175 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerStarted","Data":"002d4b36635e093b1fa011fd8aa57466323ff773fe56e2f8fd80bfc94a5305aa"} Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.223793 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2c546315-1cd0-45e3-809b-c6a8d2e509b3","Type":"ContainerStarted","Data":"893d538ffec4423a138a93f8dc4f69cc32e6080f5b2602cde0634ce61a3cd4d1"} Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.226995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" event={"ID":"bf6d608b-355b-471e-9fab-40b62a25b8f0","Type":"ContainerStarted","Data":"2ab1b4a19c367c9ffceebc495e6f631a1fd7c02935d8396230ea56f4d9c7162f"} Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.227271 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.233637 4744 generic.go:334] "Generic (PLEG): container finished" podID="992860bb-2626-401b-b4a0-439bb834f8e3" containerID="f704408826be9ed92a9d2c0dfade364364a46376f5afb28dfca7b651afcd64f2" exitCode=143 Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.233691 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"992860bb-2626-401b-b4a0-439bb834f8e3","Type":"ContainerDied","Data":"f704408826be9ed92a9d2c0dfade364364a46376f5afb28dfca7b651afcd64f2"} Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.242136 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a1b118af-7005-4a34-94ae-cf00f8c63839","Type":"ContainerStarted","Data":"6010e9d9166c24da03fda2b4eb800cfb062467531bb0a1341308cd707efc3e30"} Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.255642 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" podStartSLOduration=4.255627498 podStartE2EDuration="4.255627498s" podCreationTimestamp="2026-01-06 15:00:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:20.247464262 +0000 UTC m=+1416.874930580" watchObservedRunningTime="2026-01-06 15:00:20.255627498 +0000 UTC m=+1416.883093816" Jan 06 15:00:20 crc kubenswrapper[4744]: I0106 15:00:20.888347 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.262429 4744 generic.go:334] "Generic (PLEG): container finished" podID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerID="9281ea95ced4d12abbc3f13ffac89b4f12c76d9e8363006875383c1c108ad996" exitCode=0 Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.262600 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cab7dd39-8d15-462c-9df0-dc4df7e54db6","Type":"ContainerDied","Data":"9281ea95ced4d12abbc3f13ffac89b4f12c76d9e8363006875383c1c108ad996"} Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.265714 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a1b118af-7005-4a34-94ae-cf00f8c63839","Type":"ContainerStarted","Data":"59457c62f7da9a0f199cabf8981368dc00d8eecbaa327669e025d6b51bdcb2d7"} Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.269478 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2c546315-1cd0-45e3-809b-c6a8d2e509b3","Type":"ContainerStarted","Data":"6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9"} Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.269545 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api-log" containerID="cri-o://893d538ffec4423a138a93f8dc4f69cc32e6080f5b2602cde0634ce61a3cd4d1" gracePeriod=30 Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.269623 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api" containerID="cri-o://6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9" gracePeriod=30 Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.269729 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.299374 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.884601407 podStartE2EDuration="6.299352159s" podCreationTimestamp="2026-01-06 15:00:15 +0000 UTC" firstStartedPulling="2026-01-06 15:00:17.034862275 +0000 UTC m=+1413.662328593" lastFinishedPulling="2026-01-06 15:00:18.449613027 +0000 UTC m=+1415.077079345" observedRunningTime="2026-01-06 15:00:21.284864486 +0000 UTC m=+1417.912330814" watchObservedRunningTime="2026-01-06 15:00:21.299352159 +0000 UTC m=+1417.926818477" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.331406 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.331387486 podStartE2EDuration="5.331387486s" podCreationTimestamp="2026-01-06 15:00:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:21.310025342 +0000 UTC m=+1417.937491670" watchObservedRunningTime="2026-01-06 15:00:21.331387486 +0000 UTC m=+1417.958853804" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.340398 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.350552 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c5f6db8dc-vwrjb" podStartSLOduration=5.350532733 podStartE2EDuration="5.350532733s" podCreationTimestamp="2026-01-06 15:00:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:21.344424851 +0000 UTC m=+1417.971891169" watchObservedRunningTime="2026-01-06 15:00:21.350532733 +0000 UTC m=+1417.977999051" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.389668 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.389733 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.390038 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.390476 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.390912 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="barbican-api-log" containerStatusID={"Type":"cri-o","ID":"200eb467c785f7afde1312ce4bfe4d1dc6b36dc03264f727b5c987e731daf20e"} pod="openstack/barbican-api-7bc78458f8-mx59q" containerMessage="Container barbican-api-log failed liveness probe, will be restarted" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.390938 4744 scope.go:117] "RemoveContainer" containerID="e9ac5a9c40c477cf86db69b0841546b8eaccb75dbee02e33232bd2cfafdbcae0" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.390966 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" containerID="cri-o://200eb467c785f7afde1312ce4bfe4d1dc6b36dc03264f727b5c987e731daf20e" gracePeriod=30 Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.801603 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:21 crc kubenswrapper[4744]: I0106 15:00:21.802985 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-787d958d8c-44gnt" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.239006 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.285482 4744 generic.go:334] "Generic (PLEG): container finished" podID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerID="893d538ffec4423a138a93f8dc4f69cc32e6080f5b2602cde0634ce61a3cd4d1" exitCode=143 Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.285585 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2c546315-1cd0-45e3-809b-c6a8d2e509b3","Type":"ContainerDied","Data":"893d538ffec4423a138a93f8dc4f69cc32e6080f5b2602cde0634ce61a3cd4d1"} Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.288472 4744 generic.go:334] "Generic (PLEG): container finished" podID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerID="200eb467c785f7afde1312ce4bfe4d1dc6b36dc03264f727b5c987e731daf20e" exitCode=143 Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.288522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerDied","Data":"200eb467c785f7afde1312ce4bfe4d1dc6b36dc03264f727b5c987e731daf20e"} Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.288547 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerStarted","Data":"f0b1b2dd668aea8d6764923ae9c1bf2b53f83314e6ad7765a52b6afccbb1bc29"} Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.297351 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerStarted","Data":"bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f"} Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.567878 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.674977 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-combined-ca-bundle\") pod \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.675524 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-config-data\") pod \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.675723 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-scripts\") pod \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.675985 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-httpd-run\") pod \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.676107 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-logs\") pod \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.676129 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl82v\" (UniqueName: \"kubernetes.io/projected/cab7dd39-8d15-462c-9df0-dc4df7e54db6-kube-api-access-zl82v\") pod \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.676306 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-internal-tls-certs\") pod \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.676647 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\" (UID: \"cab7dd39-8d15-462c-9df0-dc4df7e54db6\") " Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.677716 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-logs" (OuterVolumeSpecName: "logs") pod "cab7dd39-8d15-462c-9df0-dc4df7e54db6" (UID: "cab7dd39-8d15-462c-9df0-dc4df7e54db6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.677952 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.678108 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cab7dd39-8d15-462c-9df0-dc4df7e54db6" (UID: "cab7dd39-8d15-462c-9df0-dc4df7e54db6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.685228 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cab7dd39-8d15-462c-9df0-dc4df7e54db6-kube-api-access-zl82v" (OuterVolumeSpecName: "kube-api-access-zl82v") pod "cab7dd39-8d15-462c-9df0-dc4df7e54db6" (UID: "cab7dd39-8d15-462c-9df0-dc4df7e54db6"). InnerVolumeSpecName "kube-api-access-zl82v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.706515 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b" (OuterVolumeSpecName: "glance") pod "cab7dd39-8d15-462c-9df0-dc4df7e54db6" (UID: "cab7dd39-8d15-462c-9df0-dc4df7e54db6"). InnerVolumeSpecName "pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.711706 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-scripts" (OuterVolumeSpecName: "scripts") pod "cab7dd39-8d15-462c-9df0-dc4df7e54db6" (UID: "cab7dd39-8d15-462c-9df0-dc4df7e54db6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.717348 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cab7dd39-8d15-462c-9df0-dc4df7e54db6" (UID: "cab7dd39-8d15-462c-9df0-dc4df7e54db6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.760331 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "cab7dd39-8d15-462c-9df0-dc4df7e54db6" (UID: "cab7dd39-8d15-462c-9df0-dc4df7e54db6"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.777323 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-config-data" (OuterVolumeSpecName: "config-data") pod "cab7dd39-8d15-462c-9df0-dc4df7e54db6" (UID: "cab7dd39-8d15-462c-9df0-dc4df7e54db6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.781436 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.781472 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.781483 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.781496 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cab7dd39-8d15-462c-9df0-dc4df7e54db6-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.781507 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl82v\" (UniqueName: \"kubernetes.io/projected/cab7dd39-8d15-462c-9df0-dc4df7e54db6-kube-api-access-zl82v\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.781518 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cab7dd39-8d15-462c-9df0-dc4df7e54db6-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.781543 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") on node \"crc\" " Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.784211 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b8c67fcd6-nlf9f" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.885768 4744 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.886536 4744 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b") on node "crc" Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.894799 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7bc78458f8-mx59q"] Jan 06 15:00:22 crc kubenswrapper[4744]: I0106 15:00:22.985217 4744 reconciler_common.go:293] "Volume detached for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.310463 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"cab7dd39-8d15-462c-9df0-dc4df7e54db6","Type":"ContainerDied","Data":"ea6a39d4640e2178dbdb96cef4a87983970f202703f7744008e8148e36e2d8d8"} Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.310845 4744 scope.go:117] "RemoveContainer" containerID="9281ea95ced4d12abbc3f13ffac89b4f12c76d9e8363006875383c1c108ad996" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.310482 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.340945 4744 scope.go:117] "RemoveContainer" containerID="0d55873db7667c2a58e81e8697f3b88667d3f505075d94fb72f3e82451ef5938" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.358147 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.367245 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.377086 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 15:00:23 crc kubenswrapper[4744]: E0106 15:00:23.377593 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c2a7970-7262-45b0-91c4-96b457f06692" containerName="init" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.377610 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c2a7970-7262-45b0-91c4-96b457f06692" containerName="init" Jan 06 15:00:23 crc kubenswrapper[4744]: E0106 15:00:23.377640 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerName="glance-log" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.377648 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerName="glance-log" Jan 06 15:00:23 crc kubenswrapper[4744]: E0106 15:00:23.377670 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerName="glance-httpd" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.377678 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerName="glance-httpd" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.377867 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerName="glance-log" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.377884 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c2a7970-7262-45b0-91c4-96b457f06692" containerName="init" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.377896 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" containerName="glance-httpd" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.378999 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.384275 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.384542 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.393752 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.405604 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpmwh\" (UniqueName: \"kubernetes.io/projected/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-kube-api-access-jpmwh\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.405673 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.405743 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-logs\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.405791 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.405884 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.405936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.406035 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.412346 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.508754 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.509012 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpmwh\" (UniqueName: \"kubernetes.io/projected/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-kube-api-access-jpmwh\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.509078 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.509237 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.509112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-logs\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.509557 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.509640 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.509696 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.509736 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.509780 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-logs\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.528140 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.528359 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.528588 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.528935 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.529004 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a751ba72a1be54984524758ff1fa5b9a696d9d8d73f1feb0e198d80940b6b3a7/globalmount\"" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.532606 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.533391 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpmwh\" (UniqueName: \"kubernetes.io/projected/6ae7b1ee-3df3-443f-92e9-c4a68e6829e3-kube-api-access-jpmwh\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.617668 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6520655a-b639-44d4-9ac9-c473d1b54a5b\") pod \"glance-default-internal-api-0\" (UID: \"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3\") " pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.706950 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:23 crc kubenswrapper[4744]: I0106 15:00:23.760904 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cab7dd39-8d15-462c-9df0-dc4df7e54db6" path="/var/lib/kubelet/pods/cab7dd39-8d15-462c-9df0-dc4df7e54db6/volumes" Jan 06 15:00:24 crc kubenswrapper[4744]: I0106 15:00:24.378415 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 06 15:00:25 crc kubenswrapper[4744]: I0106 15:00:25.346249 4744 generic.go:334] "Generic (PLEG): container finished" podID="992860bb-2626-401b-b4a0-439bb834f8e3" containerID="de4be668780cd37bacbeb7ad79259bca72d6ec15130dcb1a90d59bd1ee307017" exitCode=0 Jan 06 15:00:25 crc kubenswrapper[4744]: I0106 15:00:25.346552 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"992860bb-2626-401b-b4a0-439bb834f8e3","Type":"ContainerDied","Data":"de4be668780cd37bacbeb7ad79259bca72d6ec15130dcb1a90d59bd1ee307017"} Jan 06 15:00:25 crc kubenswrapper[4744]: I0106 15:00:25.347898 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3","Type":"ContainerStarted","Data":"5630cfdbe86ad494b50f42f3f4419c008861a5623e5bbabe14596d4bef15732e"} Jan 06 15:00:26 crc kubenswrapper[4744]: I0106 15:00:26.341034 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.207:8080/\": dial tcp 10.217.0.207:8080: connect: connection refused" Jan 06 15:00:26 crc kubenswrapper[4744]: I0106 15:00:26.390144 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:26 crc kubenswrapper[4744]: I0106 15:00:26.481409 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:26 crc kubenswrapper[4744]: I0106 15:00:26.576745 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-gndpz"] Jan 06 15:00:26 crc kubenswrapper[4744]: I0106 15:00:26.577265 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" podUID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" containerName="dnsmasq-dns" containerID="cri-o://c8f1d227f184e3c8f79983e9b7dd37aafb20ed8b3327dc627f8208edcd1b3001" gracePeriod=10 Jan 06 15:00:28 crc kubenswrapper[4744]: I0106 15:00:28.464866 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" podUID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.187:5353: connect: connection refused" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.059280 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.383638 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7bd5fcfc47-lf5fz"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.386331 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.393087 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-zl99k" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.393468 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.394376 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.409874 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7bd5fcfc47-lf5fz"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.441665 4744 generic.go:334] "Generic (PLEG): container finished" podID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" containerID="c8f1d227f184e3c8f79983e9b7dd37aafb20ed8b3327dc627f8208edcd1b3001" exitCode=0 Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.441780 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" event={"ID":"e39dd03e-1fc4-4650-b0b6-d0067a81ed77","Type":"ContainerDied","Data":"c8f1d227f184e3c8f79983e9b7dd37aafb20ed8b3327dc627f8208edcd1b3001"} Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.447013 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerStarted","Data":"c831bd9aa1fe0e31743d20d735e5fd47d5bd5dfea977d0c750a0018628c908be"} Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.448494 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3","Type":"ContainerStarted","Data":"6afdcb363003ca6198df166ec3b81a792018eda75afd861e3b7eb2c1bfe5269e"} Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.506440 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-sk49l"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.507926 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.507977 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data-custom\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.508073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8z9d\" (UniqueName: \"kubernetes.io/projected/13f41b35-92ba-4b8a-85cf-969aeccdba04-kube-api-access-r8z9d\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.508127 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-combined-ca-bundle\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.508207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.555742 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sk49l"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.627138 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.627705 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data-custom\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.627782 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vczd\" (UniqueName: \"kubernetes.io/projected/1174ed04-5464-450f-850f-e77c11afee5d-kube-api-access-5vczd\") pod \"nova-api-db-create-sk49l\" (UID: \"1174ed04-5464-450f-850f-e77c11afee5d\") " pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.627894 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8z9d\" (UniqueName: \"kubernetes.io/projected/13f41b35-92ba-4b8a-85cf-969aeccdba04-kube-api-access-r8z9d\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.627989 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1174ed04-5464-450f-850f-e77c11afee5d-operator-scripts\") pod \"nova-api-db-create-sk49l\" (UID: \"1174ed04-5464-450f-850f-e77c11afee5d\") " pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.628022 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-combined-ca-bundle\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.671039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data-custom\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.681934 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8z9d\" (UniqueName: \"kubernetes.io/projected/13f41b35-92ba-4b8a-85cf-969aeccdba04-kube-api-access-r8z9d\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.690258 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.755701 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-combined-ca-bundle\") pod \"heat-engine-7bd5fcfc47-lf5fz\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.821137 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1174ed04-5464-450f-850f-e77c11afee5d-operator-scripts\") pod \"nova-api-db-create-sk49l\" (UID: \"1174ed04-5464-450f-850f-e77c11afee5d\") " pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.821697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vczd\" (UniqueName: \"kubernetes.io/projected/1174ed04-5464-450f-850f-e77c11afee5d-kube-api-access-5vczd\") pod \"nova-api-db-create-sk49l\" (UID: \"1174ed04-5464-450f-850f-e77c11afee5d\") " pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.824548 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1174ed04-5464-450f-850f-e77c11afee5d-operator-scripts\") pod \"nova-api-db-create-sk49l\" (UID: \"1174ed04-5464-450f-850f-e77c11afee5d\") " pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.867569 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jvkwj"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.873049 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.881464 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vczd\" (UniqueName: \"kubernetes.io/projected/1174ed04-5464-450f-850f-e77c11afee5d-kube-api-access-5vczd\") pod \"nova-api-db-create-sk49l\" (UID: \"1174ed04-5464-450f-850f-e77c11afee5d\") " pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.894296 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jvkwj"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.915150 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-b7ad-account-create-update-s4ph8"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.916961 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.922079 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.953547 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b7ad-account-create-update-s4ph8"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.955298 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wwj9s"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.958145 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.976225 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-69c8df499f-wjrkv"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.977729 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.986050 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-64557f6b9b-kxvsw"] Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.987136 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.987879 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:30 crc kubenswrapper[4744]: I0106 15:00:30.989328 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.002105 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wwj9s"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.011366 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69c8df499f-wjrkv"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.030275 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-8jcdj"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.031599 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.032304 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.040205 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-64557f6b9b-kxvsw"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.052889 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-8jcdj"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.054468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv59j\" (UniqueName: \"kubernetes.io/projected/49c698d4-2488-47f1-ad67-7506ec912b38-kube-api-access-xv59j\") pod \"nova-cell0-db-create-jvkwj\" (UID: \"49c698d4-2488-47f1-ad67-7506ec912b38\") " pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.054522 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49c698d4-2488-47f1-ad67-7506ec912b38-operator-scripts\") pod \"nova-cell0-db-create-jvkwj\" (UID: \"49c698d4-2488-47f1-ad67-7506ec912b38\") " pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.054552 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57295aa-50a7-478a-805c-a682bcf0386b-operator-scripts\") pod \"nova-api-b7ad-account-create-update-s4ph8\" (UID: \"e57295aa-50a7-478a-805c-a682bcf0386b\") " pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.054586 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9pk6\" (UniqueName: \"kubernetes.io/projected/e57295aa-50a7-478a-805c-a682bcf0386b-kube-api-access-s9pk6\") pod \"nova-api-b7ad-account-create-update-s4ph8\" (UID: \"e57295aa-50a7-478a-805c-a682bcf0386b\") " pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.073694 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-63fc-account-create-update-bjlhb"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.075274 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.088457 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.093526 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-63fc-account-create-update-bjlhb"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.128471 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.145348 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-afb3-account-create-update-k5htr"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.147397 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.152522 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156073 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156103 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v677j\" (UniqueName: \"kubernetes.io/projected/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-kube-api-access-v677j\") pod \"nova-cell1-db-create-8jcdj\" (UID: \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\") " pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156154 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq56h\" (UniqueName: \"kubernetes.io/projected/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-kube-api-access-hq56h\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156190 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-operator-scripts\") pod \"nova-cell1-db-create-8jcdj\" (UID: \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\") " pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156222 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data-custom\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156255 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156275 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data-custom\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156304 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv59j\" (UniqueName: \"kubernetes.io/projected/49c698d4-2488-47f1-ad67-7506ec912b38-kube-api-access-xv59j\") pod \"nova-cell0-db-create-jvkwj\" (UID: \"49c698d4-2488-47f1-ad67-7506ec912b38\") " pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156329 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49c698d4-2488-47f1-ad67-7506ec912b38-operator-scripts\") pod \"nova-cell0-db-create-jvkwj\" (UID: \"49c698d4-2488-47f1-ad67-7506ec912b38\") " pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57295aa-50a7-478a-805c-a682bcf0386b-operator-scripts\") pod \"nova-api-b7ad-account-create-update-s4ph8\" (UID: \"e57295aa-50a7-478a-805c-a682bcf0386b\") " pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156374 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9pk6\" (UniqueName: \"kubernetes.io/projected/e57295aa-50a7-478a-805c-a682bcf0386b-kube-api-access-s9pk6\") pod \"nova-api-b7ad-account-create-update-s4ph8\" (UID: \"e57295aa-50a7-478a-805c-a682bcf0386b\") " pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156401 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-combined-ca-bundle\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156419 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156434 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb26c\" (UniqueName: \"kubernetes.io/projected/f1e405ef-b09a-47b9-902e-db0ff2df7788-kube-api-access-cb26c\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156453 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-config\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156474 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156493 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156528 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-combined-ca-bundle\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156550 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm526\" (UniqueName: \"kubernetes.io/projected/aefa115f-c019-4d47-be4c-659e19fe9eea-kube-api-access-dm526\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.156574 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-svc\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.164866 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49c698d4-2488-47f1-ad67-7506ec912b38-operator-scripts\") pod \"nova-cell0-db-create-jvkwj\" (UID: \"49c698d4-2488-47f1-ad67-7506ec912b38\") " pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.166084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57295aa-50a7-478a-805c-a682bcf0386b-operator-scripts\") pod \"nova-api-b7ad-account-create-update-s4ph8\" (UID: \"e57295aa-50a7-478a-805c-a682bcf0386b\") " pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.168531 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-afb3-account-create-update-k5htr"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.186892 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv59j\" (UniqueName: \"kubernetes.io/projected/49c698d4-2488-47f1-ad67-7506ec912b38-kube-api-access-xv59j\") pod \"nova-cell0-db-create-jvkwj\" (UID: \"49c698d4-2488-47f1-ad67-7506ec912b38\") " pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.194450 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9pk6\" (UniqueName: \"kubernetes.io/projected/e57295aa-50a7-478a-805c-a682bcf0386b-kube-api-access-s9pk6\") pod \"nova-api-b7ad-account-create-update-s4ph8\" (UID: \"e57295aa-50a7-478a-805c-a682bcf0386b\") " pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-svc\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258675 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258700 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v677j\" (UniqueName: \"kubernetes.io/projected/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-kube-api-access-v677j\") pod \"nova-cell1-db-create-8jcdj\" (UID: \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\") " pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258730 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5n6z\" (UniqueName: \"kubernetes.io/projected/49796ee8-5c97-4c55-a8d8-4c47279d2641-kube-api-access-j5n6z\") pod \"nova-cell0-63fc-account-create-update-bjlhb\" (UID: \"49796ee8-5c97-4c55-a8d8-4c47279d2641\") " pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258773 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49796ee8-5c97-4c55-a8d8-4c47279d2641-operator-scripts\") pod \"nova-cell0-63fc-account-create-update-bjlhb\" (UID: \"49796ee8-5c97-4c55-a8d8-4c47279d2641\") " pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258793 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq56h\" (UniqueName: \"kubernetes.io/projected/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-kube-api-access-hq56h\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258817 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-operator-scripts\") pod \"nova-cell1-db-create-8jcdj\" (UID: \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\") " pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258853 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data-custom\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258887 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258912 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data-custom\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258933 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdtx7\" (UniqueName: \"kubernetes.io/projected/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-kube-api-access-xdtx7\") pod \"nova-cell1-afb3-account-create-update-k5htr\" (UID: \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\") " pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.258995 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-combined-ca-bundle\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.259014 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.259032 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb26c\" (UniqueName: \"kubernetes.io/projected/f1e405ef-b09a-47b9-902e-db0ff2df7788-kube-api-access-cb26c\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.259055 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-operator-scripts\") pod \"nova-cell1-afb3-account-create-update-k5htr\" (UID: \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\") " pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.259074 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-config\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.259097 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.259175 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.259224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-combined-ca-bundle\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.259246 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm526\" (UniqueName: \"kubernetes.io/projected/aefa115f-c019-4d47-be4c-659e19fe9eea-kube-api-access-dm526\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.259944 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-swift-storage-0\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.260722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-operator-scripts\") pod \"nova-cell1-db-create-8jcdj\" (UID: \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\") " pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.260754 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-svc\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.263609 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-nb\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.264602 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-sb\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.267013 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-config\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.272948 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.274983 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data-custom\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.276765 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data-custom\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.290974 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.301402 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.303078 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-combined-ca-bundle\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.304785 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb26c\" (UniqueName: \"kubernetes.io/projected/f1e405ef-b09a-47b9-902e-db0ff2df7788-kube-api-access-cb26c\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.305953 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-combined-ca-bundle\") pod \"heat-cfnapi-64557f6b9b-kxvsw\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.308883 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm526\" (UniqueName: \"kubernetes.io/projected/aefa115f-c019-4d47-be4c-659e19fe9eea-kube-api-access-dm526\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.310438 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data\") pod \"heat-api-69c8df499f-wjrkv\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.316995 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.319707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v677j\" (UniqueName: \"kubernetes.io/projected/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-kube-api-access-v677j\") pod \"nova-cell1-db-create-8jcdj\" (UID: \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\") " pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.347409 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq56h\" (UniqueName: \"kubernetes.io/projected/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-kube-api-access-hq56h\") pod \"dnsmasq-dns-7d978555f9-wwj9s\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.356402 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.360764 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sztt\" (UniqueName: \"kubernetes.io/projected/992860bb-2626-401b-b4a0-439bb834f8e3-kube-api-access-4sztt\") pod \"992860bb-2626-401b-b4a0-439bb834f8e3\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.360872 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-combined-ca-bundle\") pod \"992860bb-2626-401b-b4a0-439bb834f8e3\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.360928 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-logs\") pod \"992860bb-2626-401b-b4a0-439bb834f8e3\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.360944 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-scripts\") pod \"992860bb-2626-401b-b4a0-439bb834f8e3\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.361001 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-config-data\") pod \"992860bb-2626-401b-b4a0-439bb834f8e3\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.361021 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-public-tls-certs\") pod \"992860bb-2626-401b-b4a0-439bb834f8e3\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.361051 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-httpd-run\") pod \"992860bb-2626-401b-b4a0-439bb834f8e3\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.361261 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"992860bb-2626-401b-b4a0-439bb834f8e3\" (UID: \"992860bb-2626-401b-b4a0-439bb834f8e3\") " Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.361576 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5n6z\" (UniqueName: \"kubernetes.io/projected/49796ee8-5c97-4c55-a8d8-4c47279d2641-kube-api-access-j5n6z\") pod \"nova-cell0-63fc-account-create-update-bjlhb\" (UID: \"49796ee8-5c97-4c55-a8d8-4c47279d2641\") " pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.361625 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49796ee8-5c97-4c55-a8d8-4c47279d2641-operator-scripts\") pod \"nova-cell0-63fc-account-create-update-bjlhb\" (UID: \"49796ee8-5c97-4c55-a8d8-4c47279d2641\") " pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.361716 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdtx7\" (UniqueName: \"kubernetes.io/projected/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-kube-api-access-xdtx7\") pod \"nova-cell1-afb3-account-create-update-k5htr\" (UID: \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\") " pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.361784 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-operator-scripts\") pod \"nova-cell1-afb3-account-create-update-k5htr\" (UID: \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\") " pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.362495 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-operator-scripts\") pod \"nova-cell1-afb3-account-create-update-k5htr\" (UID: \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\") " pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.362707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49796ee8-5c97-4c55-a8d8-4c47279d2641-operator-scripts\") pod \"nova-cell0-63fc-account-create-update-bjlhb\" (UID: \"49796ee8-5c97-4c55-a8d8-4c47279d2641\") " pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.364562 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "992860bb-2626-401b-b4a0-439bb834f8e3" (UID: "992860bb-2626-401b-b4a0-439bb834f8e3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.364965 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-logs" (OuterVolumeSpecName: "logs") pod "992860bb-2626-401b-b4a0-439bb834f8e3" (UID: "992860bb-2626-401b-b4a0-439bb834f8e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.370419 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-scripts" (OuterVolumeSpecName: "scripts") pod "992860bb-2626-401b-b4a0-439bb834f8e3" (UID: "992860bb-2626-401b-b4a0-439bb834f8e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.380072 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5n6z\" (UniqueName: \"kubernetes.io/projected/49796ee8-5c97-4c55-a8d8-4c47279d2641-kube-api-access-j5n6z\") pod \"nova-cell0-63fc-account-create-update-bjlhb\" (UID: \"49796ee8-5c97-4c55-a8d8-4c47279d2641\") " pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.383781 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdtx7\" (UniqueName: \"kubernetes.io/projected/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-kube-api-access-xdtx7\") pod \"nova-cell1-afb3-account-create-update-k5htr\" (UID: \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\") " pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.391956 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.392129 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.427697 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "992860bb-2626-401b-b4a0-439bb834f8e3" (UID: "992860bb-2626-401b-b4a0-439bb834f8e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.428511 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.432056 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0" (OuterVolumeSpecName: "glance") pod "992860bb-2626-401b-b4a0-439bb834f8e3" (UID: "992860bb-2626-401b-b4a0-439bb834f8e3"). InnerVolumeSpecName "pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.447855 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "992860bb-2626-401b-b4a0-439bb834f8e3" (UID: "992860bb-2626-401b-b4a0-439bb834f8e3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.457302 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.463554 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.463598 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.463610 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.463623 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/992860bb-2626-401b-b4a0-439bb834f8e3-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.463669 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") on node \"crc\" " Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.463689 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.482029 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.506134 4744 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.506220 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.506318 4744 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0") on node "crc" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.511390 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/992860bb-2626-401b-b4a0-439bb834f8e3-kube-api-access-4sztt" (OuterVolumeSpecName: "kube-api-access-4sztt") pod "992860bb-2626-401b-b4a0-439bb834f8e3" (UID: "992860bb-2626-401b-b4a0-439bb834f8e3"). InnerVolumeSpecName "kube-api-access-4sztt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.522299 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-config-data" (OuterVolumeSpecName: "config-data") pod "992860bb-2626-401b-b4a0-439bb834f8e3" (UID: "992860bb-2626-401b-b4a0-439bb834f8e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.551557 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.551546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"992860bb-2626-401b-b4a0-439bb834f8e3","Type":"ContainerDied","Data":"05de888a323b4e236b1dec49054d625def33bbb0f01d2c8fa315cf9128697c32"} Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.551749 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" containerID="cri-o://c831bd9aa1fe0e31743d20d735e5fd47d5bd5dfea977d0c750a0018628c908be" gracePeriod=30 Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.551900 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.551967 4744 scope.go:117] "RemoveContainer" containerID="de4be668780cd37bacbeb7ad79259bca72d6ec15130dcb1a90d59bd1ee307017" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.551694 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" containerID="cri-o://f0b1b2dd668aea8d6764923ae9c1bf2b53f83314e6ad7765a52b6afccbb1bc29" gracePeriod=30 Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.552132 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.558766 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7bc78458f8-mx59q" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": dial tcp 10.217.0.202:9311: connect: connection refused" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.567703 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992860bb-2626-401b-b4a0-439bb834f8e3-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.567730 4744 reconciler_common.go:293] "Volume detached for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.567746 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sztt\" (UniqueName: \"kubernetes.io/projected/992860bb-2626-401b-b4a0-439bb834f8e3-kube-api-access-4sztt\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.647203 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.786437 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.786496 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.912995 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 15:00:31 crc kubenswrapper[4744]: E0106 15:00:31.915592 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992860bb-2626-401b-b4a0-439bb834f8e3" containerName="glance-httpd" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.915621 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="992860bb-2626-401b-b4a0-439bb834f8e3" containerName="glance-httpd" Jan 06 15:00:31 crc kubenswrapper[4744]: E0106 15:00:31.915635 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992860bb-2626-401b-b4a0-439bb834f8e3" containerName="glance-log" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.915642 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="992860bb-2626-401b-b4a0-439bb834f8e3" containerName="glance-log" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.915895 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="992860bb-2626-401b-b4a0-439bb834f8e3" containerName="glance-httpd" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.915922 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="992860bb-2626-401b-b4a0-439bb834f8e3" containerName="glance-log" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.917717 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.953121 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.953363 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 06 15:00:31 crc kubenswrapper[4744]: I0106 15:00:31.975314 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.015468 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.026280 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-sk49l"] Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.039389 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7bd5fcfc47-lf5fz"] Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.044148 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.044220 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.044257 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef75f535-8a96-40b4-9719-965a1c97a9be-logs\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.044280 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-scripts\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.044317 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx5tc\" (UniqueName: \"kubernetes.io/projected/ef75f535-8a96-40b4-9719-965a1c97a9be-kube-api-access-hx5tc\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.044344 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.044390 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-config-data\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.044614 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef75f535-8a96-40b4-9719-965a1c97a9be-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.150035 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-config-data\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.150408 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef75f535-8a96-40b4-9719-965a1c97a9be-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.150513 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.150533 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.150765 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef75f535-8a96-40b4-9719-965a1c97a9be-logs\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.150788 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-scripts\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.150866 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx5tc\" (UniqueName: \"kubernetes.io/projected/ef75f535-8a96-40b4-9719-965a1c97a9be-kube-api-access-hx5tc\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.150893 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.151127 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef75f535-8a96-40b4-9719-965a1c97a9be-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.151619 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef75f535-8a96-40b4-9719-965a1c97a9be-logs\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.158081 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-scripts\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.160707 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.162280 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.162348 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.162373 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ac0041d382eb549e2b0d342e4bfe43b8e025e345c53521653e27358ce8689034/globalmount\"" pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.167321 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef75f535-8a96-40b4-9719-965a1c97a9be-config-data\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.205837 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx5tc\" (UniqueName: \"kubernetes.io/projected/ef75f535-8a96-40b4-9719-965a1c97a9be-kube-api-access-hx5tc\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.228699 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6fc95a5-d19e-4b91-954f-35630f3c8ba0\") pod \"glance-default-external-api-0\" (UID: \"ef75f535-8a96-40b4-9719-965a1c97a9be\") " pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.274806 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 06 15:00:32 crc kubenswrapper[4744]: W0106 15:00:32.493492 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1174ed04_5464_450f_850f_e77c11afee5d.slice/crio-59e4e148b4098deb90f9714997720e30e12f688d9a6199995d482509e33b14b9 WatchSource:0}: Error finding container 59e4e148b4098deb90f9714997720e30e12f688d9a6199995d482509e33b14b9: Status 404 returned error can't find the container with id 59e4e148b4098deb90f9714997720e30e12f688d9a6199995d482509e33b14b9 Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.497549 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jvkwj"] Jan 06 15:00:32 crc kubenswrapper[4744]: W0106 15:00:32.506045 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice/crio-b890912e39343e60abdcfe8bf6b589efc3c4b41a586c0bccc7424b18c53a8ef3 WatchSource:0}: Error finding container b890912e39343e60abdcfe8bf6b589efc3c4b41a586c0bccc7424b18c53a8ef3: Status 404 returned error can't find the container with id b890912e39343e60abdcfe8bf6b589efc3c4b41a586c0bccc7424b18c53a8ef3 Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.586464 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" event={"ID":"13f41b35-92ba-4b8a-85cf-969aeccdba04","Type":"ContainerStarted","Data":"b890912e39343e60abdcfe8bf6b589efc3c4b41a586c0bccc7424b18c53a8ef3"} Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.599227 4744 generic.go:334] "Generic (PLEG): container finished" podID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerID="c831bd9aa1fe0e31743d20d735e5fd47d5bd5dfea977d0c750a0018628c908be" exitCode=1 Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.599259 4744 generic.go:334] "Generic (PLEG): container finished" podID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerID="f0b1b2dd668aea8d6764923ae9c1bf2b53f83314e6ad7765a52b6afccbb1bc29" exitCode=143 Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.599303 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerDied","Data":"c831bd9aa1fe0e31743d20d735e5fd47d5bd5dfea977d0c750a0018628c908be"} Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.599330 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerDied","Data":"f0b1b2dd668aea8d6764923ae9c1bf2b53f83314e6ad7765a52b6afccbb1bc29"} Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.600925 4744 scope.go:117] "RemoveContainer" containerID="f704408826be9ed92a9d2c0dfade364364a46376f5afb28dfca7b651afcd64f2" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.611111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6ae7b1ee-3df3-443f-92e9-c4a68e6829e3","Type":"ContainerStarted","Data":"b5dec3d73555cf04f8cd0045e42082e0662fe0c7ff51a4efc62434bf8928353d"} Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.613346 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerName="cinder-scheduler" containerID="cri-o://6010e9d9166c24da03fda2b4eb800cfb062467531bb0a1341308cd707efc3e30" gracePeriod=30 Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.613469 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sk49l" event={"ID":"1174ed04-5464-450f-850f-e77c11afee5d","Type":"ContainerStarted","Data":"59e4e148b4098deb90f9714997720e30e12f688d9a6199995d482509e33b14b9"} Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.613524 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerName="probe" containerID="cri-o://59457c62f7da9a0f199cabf8981368dc00d8eecbaa327669e025d6b51bdcb2d7" gracePeriod=30 Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.661801 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.675082 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.675059236 podStartE2EDuration="9.675059236s" podCreationTimestamp="2026-01-06 15:00:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:32.642293789 +0000 UTC m=+1429.269760107" watchObservedRunningTime="2026-01-06 15:00:32.675059236 +0000 UTC m=+1429.302525554" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.767995 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-svc\") pod \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.768085 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-config\") pod \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.772847 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28xgr\" (UniqueName: \"kubernetes.io/projected/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-kube-api-access-28xgr\") pod \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.772897 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-sb\") pod \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.773007 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-nb\") pod \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.773094 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-swift-storage-0\") pod \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\" (UID: \"e39dd03e-1fc4-4650-b0b6-d0067a81ed77\") " Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.860477 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-kube-api-access-28xgr" (OuterVolumeSpecName: "kube-api-access-28xgr") pod "e39dd03e-1fc4-4650-b0b6-d0067a81ed77" (UID: "e39dd03e-1fc4-4650-b0b6-d0067a81ed77"). InnerVolumeSpecName "kube-api-access-28xgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:32 crc kubenswrapper[4744]: I0106 15:00:32.877786 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28xgr\" (UniqueName: \"kubernetes.io/projected/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-kube-api-access-28xgr\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.006031 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e39dd03e-1fc4-4650-b0b6-d0067a81ed77" (UID: "e39dd03e-1fc4-4650-b0b6-d0067a81ed77"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.066381 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e39dd03e-1fc4-4650-b0b6-d0067a81ed77" (UID: "e39dd03e-1fc4-4650-b0b6-d0067a81ed77"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.086629 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e39dd03e-1fc4-4650-b0b6-d0067a81ed77" (UID: "e39dd03e-1fc4-4650-b0b6-d0067a81ed77"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.089567 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.089595 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.089606 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.114864 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-config" (OuterVolumeSpecName: "config") pod "e39dd03e-1fc4-4650-b0b6-d0067a81ed77" (UID: "e39dd03e-1fc4-4650-b0b6-d0067a81ed77"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.162797 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e39dd03e-1fc4-4650-b0b6-d0067a81ed77" (UID: "e39dd03e-1fc4-4650-b0b6-d0067a81ed77"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.182139 4744 scope.go:117] "RemoveContainer" containerID="e9ac5a9c40c477cf86db69b0841546b8eaccb75dbee02e33232bd2cfafdbcae0" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.191825 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.192150 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e39dd03e-1fc4-4650-b0b6-d0067a81ed77-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.517840 4744 scope.go:117] "RemoveContainer" containerID="200eb467c785f7afde1312ce4bfe4d1dc6b36dc03264f727b5c987e731daf20e" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.644042 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b7ad-account-create-update-s4ph8"] Jan 06 15:00:33 crc kubenswrapper[4744]: W0106 15:00:33.670806 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode57295aa_50a7_478a_805c_a682bcf0386b.slice/crio-7a70955ab0f4ca2944634c7bc9f8151d51550c861344c6a5cdde656727359fd6 WatchSource:0}: Error finding container 7a70955ab0f4ca2944634c7bc9f8151d51550c861344c6a5cdde656727359fd6: Status 404 returned error can't find the container with id 7a70955ab0f4ca2944634c7bc9f8151d51550c861344c6a5cdde656727359fd6 Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.707298 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.707339 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.723379 4744 generic.go:334] "Generic (PLEG): container finished" podID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerID="59457c62f7da9a0f199cabf8981368dc00d8eecbaa327669e025d6b51bdcb2d7" exitCode=0 Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.758203 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.794231 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="992860bb-2626-401b-b4a0-439bb834f8e3" path="/var/lib/kubelet/pods/992860bb-2626-401b-b4a0-439bb834f8e3/volumes" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.795379 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a1b118af-7005-4a34-94ae-cf00f8c63839","Type":"ContainerDied","Data":"59457c62f7da9a0f199cabf8981368dc00d8eecbaa327669e025d6b51bdcb2d7"} Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.795405 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" event={"ID":"e39dd03e-1fc4-4650-b0b6-d0067a81ed77","Type":"ContainerDied","Data":"9a0de90f94af9faad3fed0d2239bbf58f5c9b3dc9346c05b9fe896d3f8611b3b"} Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.843287 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jvkwj" event={"ID":"49c698d4-2488-47f1-ad67-7506ec912b38","Type":"ContainerStarted","Data":"68a6fd9ddca2398e4f52fc5730cd5948836f9b5c460687af0873f8b0fbeb8285"} Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.917424 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:33 crc kubenswrapper[4744]: I0106 15:00:33.929487 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.061045 4744 scope.go:117] "RemoveContainer" containerID="c8f1d227f184e3c8f79983e9b7dd37aafb20ed8b3327dc627f8208edcd1b3001" Jan 06 15:00:34 crc kubenswrapper[4744]: E0106 15:00:34.141929 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1b118af_7005_4a34_94ae_cf00f8c63839.slice/crio-conmon-59457c62f7da9a0f199cabf8981368dc00d8eecbaa327669e025d6b51bdcb2d7.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.146371 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.171419 4744 scope.go:117] "RemoveContainer" containerID="0f83d8110bf3bd2c85ca8494886b4c9d3b38d50825e4441c41bdc97998490136" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.275182 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-combined-ca-bundle\") pod \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.275299 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data\") pod \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.275429 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-logs\") pod \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.275498 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data-custom\") pod \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.275635 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tn6cb\" (UniqueName: \"kubernetes.io/projected/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-kube-api-access-tn6cb\") pod \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\" (UID: \"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65\") " Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.279087 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-logs" (OuterVolumeSpecName: "logs") pod "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" (UID: "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.284639 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-kube-api-access-tn6cb" (OuterVolumeSpecName: "kube-api-access-tn6cb") pod "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" (UID: "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65"). InnerVolumeSpecName "kube-api-access-tn6cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.295630 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" (UID: "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.378426 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.378463 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.378480 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tn6cb\" (UniqueName: \"kubernetes.io/projected/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-kube-api-access-tn6cb\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.404426 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-63fc-account-create-update-bjlhb"] Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.463034 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wwj9s"] Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.501470 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-8jcdj"] Jan 06 15:00:34 crc kubenswrapper[4744]: W0106 15:00:34.519445 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d1c4236_2610_4bcd_91f7_c157f5da7ea4.slice/crio-d6568f22b1f89e404ca83ba13ff0b35f5885680e2287d4c4739ac10a55f6da8e WatchSource:0}: Error finding container d6568f22b1f89e404ca83ba13ff0b35f5885680e2287d4c4739ac10a55f6da8e: Status 404 returned error can't find the container with id d6568f22b1f89e404ca83ba13ff0b35f5885680e2287d4c4739ac10a55f6da8e Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.626563 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" (UID: "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.692456 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.806262 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data" (OuterVolumeSpecName: "config-data") pod "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" (UID: "8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.897146 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.914997 4744 generic.go:334] "Generic (PLEG): container finished" podID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerID="6010e9d9166c24da03fda2b4eb800cfb062467531bb0a1341308cd707efc3e30" exitCode=0 Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.915050 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a1b118af-7005-4a34-94ae-cf00f8c63839","Type":"ContainerDied","Data":"6010e9d9166c24da03fda2b4eb800cfb062467531bb0a1341308cd707efc3e30"} Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.943984 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7bc78458f8-mx59q" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.944806 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7bc78458f8-mx59q" event={"ID":"8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65","Type":"ContainerDied","Data":"5cf8174ae0a171b6d6168b2471644caa3df40634fbcfa94b56fc0fd14dbdc4a9"} Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.944845 4744 scope.go:117] "RemoveContainer" containerID="c831bd9aa1fe0e31743d20d735e5fd47d5bd5dfea977d0c750a0018628c908be" Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.969720 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b7ad-account-create-update-s4ph8" event={"ID":"e57295aa-50a7-478a-805c-a682bcf0386b","Type":"ContainerStarted","Data":"2f2600111ba19818cd3e8825d0359df3c02a05a77cdbe33bd933478b8cd73829"} Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.969774 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b7ad-account-create-update-s4ph8" event={"ID":"e57295aa-50a7-478a-805c-a682bcf0386b","Type":"ContainerStarted","Data":"7a70955ab0f4ca2944634c7bc9f8151d51550c861344c6a5cdde656727359fd6"} Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.983680 4744 generic.go:334] "Generic (PLEG): container finished" podID="1174ed04-5464-450f-850f-e77c11afee5d" containerID="b7bdb3b47955e4488022736ca45ed699e39c0136c090450314114a05683ad88e" exitCode=0 Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.983744 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sk49l" event={"ID":"1174ed04-5464-450f-850f-e77c11afee5d","Type":"ContainerDied","Data":"b7bdb3b47955e4488022736ca45ed699e39c0136c090450314114a05683ad88e"} Jan 06 15:00:34 crc kubenswrapper[4744]: I0106 15:00:34.986177 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" event={"ID":"49796ee8-5c97-4c55-a8d8-4c47279d2641","Type":"ContainerStarted","Data":"3c99a231b9c99589b6a22f64c8560401e0f398548f37754aa102a4c2b373722b"} Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.000814 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" event={"ID":"13f41b35-92ba-4b8a-85cf-969aeccdba04","Type":"ContainerStarted","Data":"a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6"} Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.005788 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.014429 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" event={"ID":"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a","Type":"ContainerStarted","Data":"5d1bff51a8fbd687e3e0365a6cf19387b2dc6630f066f75fad402b9d88db9180"} Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.014840 4744 scope.go:117] "RemoveContainer" containerID="f0b1b2dd668aea8d6764923ae9c1bf2b53f83314e6ad7765a52b6afccbb1bc29" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.033263 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69c8df499f-wjrkv"] Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.033333 4744 generic.go:334] "Generic (PLEG): container finished" podID="49c698d4-2488-47f1-ad67-7506ec912b38" containerID="032bc9033f646fb0f2a23fc2d0462ef4c3fdb4b022205c7168d5783f751191c4" exitCode=0 Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.033485 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jvkwj" event={"ID":"49c698d4-2488-47f1-ad67-7506ec912b38","Type":"ContainerDied","Data":"032bc9033f646fb0f2a23fc2d0462ef4c3fdb4b022205c7168d5783f751191c4"} Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.036995 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-b7ad-account-create-update-s4ph8" podStartSLOduration=5.036974263 podStartE2EDuration="5.036974263s" podCreationTimestamp="2026-01-06 15:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:34.991620423 +0000 UTC m=+1431.619086761" watchObservedRunningTime="2026-01-06 15:00:35.036974263 +0000 UTC m=+1431.664440581" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.037968 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8jcdj" event={"ID":"6d1c4236-2610-4bcd-91f7-c157f5da7ea4","Type":"ContainerStarted","Data":"d6568f22b1f89e404ca83ba13ff0b35f5885680e2287d4c4739ac10a55f6da8e"} Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.060855 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerStarted","Data":"67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9"} Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.060894 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.060915 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.086675 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-afb3-account-create-update-k5htr"] Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.121225 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-64557f6b9b-kxvsw"] Jan 06 15:00:35 crc kubenswrapper[4744]: W0106 15:00:35.133053 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-ed2df8acb6444d1b71421733b7082c76977cdb3527c193e5cd9910c5c9dbb63c WatchSource:0}: Error finding container ed2df8acb6444d1b71421733b7082c76977cdb3527c193e5cd9910c5c9dbb63c: Status 404 returned error can't find the container with id ed2df8acb6444d1b71421733b7082c76977cdb3527c193e5cd9910c5c9dbb63c Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.164671 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" podStartSLOduration=5.164648543 podStartE2EDuration="5.164648543s" podCreationTimestamp="2026-01-06 15:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:35.032904375 +0000 UTC m=+1431.660370694" watchObservedRunningTime="2026-01-06 15:00:35.164648543 +0000 UTC m=+1431.792114861" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.202858 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7bc78458f8-mx59q"] Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.228956 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7bc78458f8-mx59q"] Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.298710 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.418948 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1b118af-7005-4a34-94ae-cf00f8c63839-etc-machine-id\") pod \"a1b118af-7005-4a34-94ae-cf00f8c63839\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.419067 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data\") pod \"a1b118af-7005-4a34-94ae-cf00f8c63839\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.419186 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-scripts\") pod \"a1b118af-7005-4a34-94ae-cf00f8c63839\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.419304 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-combined-ca-bundle\") pod \"a1b118af-7005-4a34-94ae-cf00f8c63839\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.419354 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px7nn\" (UniqueName: \"kubernetes.io/projected/a1b118af-7005-4a34-94ae-cf00f8c63839-kube-api-access-px7nn\") pod \"a1b118af-7005-4a34-94ae-cf00f8c63839\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.419411 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data-custom\") pod \"a1b118af-7005-4a34-94ae-cf00f8c63839\" (UID: \"a1b118af-7005-4a34-94ae-cf00f8c63839\") " Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.427367 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-scripts" (OuterVolumeSpecName: "scripts") pod "a1b118af-7005-4a34-94ae-cf00f8c63839" (UID: "a1b118af-7005-4a34-94ae-cf00f8c63839"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.427606 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1b118af-7005-4a34-94ae-cf00f8c63839-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a1b118af-7005-4a34-94ae-cf00f8c63839" (UID: "a1b118af-7005-4a34-94ae-cf00f8c63839"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.431271 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a1b118af-7005-4a34-94ae-cf00f8c63839" (UID: "a1b118af-7005-4a34-94ae-cf00f8c63839"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.436787 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1b118af-7005-4a34-94ae-cf00f8c63839-kube-api-access-px7nn" (OuterVolumeSpecName: "kube-api-access-px7nn") pod "a1b118af-7005-4a34-94ae-cf00f8c63839" (UID: "a1b118af-7005-4a34-94ae-cf00f8c63839"). InnerVolumeSpecName "kube-api-access-px7nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:35 crc kubenswrapper[4744]: W0106 15:00:35.448742 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef75f535_8a96_40b4_9719_965a1c97a9be.slice/crio-de1e6e98320617a3e5b479d4b7a7572afabbab97e17ab18491f5a929d40b7f25 WatchSource:0}: Error finding container de1e6e98320617a3e5b479d4b7a7572afabbab97e17ab18491f5a929d40b7f25: Status 404 returned error can't find the container with id de1e6e98320617a3e5b479d4b7a7572afabbab97e17ab18491f5a929d40b7f25 Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.503234 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.522530 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px7nn\" (UniqueName: \"kubernetes.io/projected/a1b118af-7005-4a34-94ae-cf00f8c63839-kube-api-access-px7nn\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.522560 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.522569 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a1b118af-7005-4a34-94ae-cf00f8c63839-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.522579 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.577094 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1b118af-7005-4a34-94ae-cf00f8c63839" (UID: "a1b118af-7005-4a34-94ae-cf00f8c63839"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.629585 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.720669 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data" (OuterVolumeSpecName: "config-data") pod "a1b118af-7005-4a34-94ae-cf00f8c63839" (UID: "a1b118af-7005-4a34-94ae-cf00f8c63839"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.740739 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b118af-7005-4a34-94ae-cf00f8c63839-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:35 crc kubenswrapper[4744]: I0106 15:00:35.752567 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" path="/var/lib/kubelet/pods/8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65/volumes" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.089806 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" containerID="1470ffa3a8a033c5e77e522e47d32fdb0464af0fd74f285dd226ee223696052b" exitCode=0 Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.089909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" event={"ID":"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a","Type":"ContainerDied","Data":"1470ffa3a8a033c5e77e522e47d32fdb0464af0fd74f285dd226ee223696052b"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.109550 4744 generic.go:334] "Generic (PLEG): container finished" podID="c2afd989-96a7-4562-9ffd-1fac6a8f3acd" containerID="d2ba9c57984ae211313cabc1cac7e28a4fc4db714de0a470aaa2cc18bf0e8792" exitCode=0 Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.109643 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-afb3-account-create-update-k5htr" event={"ID":"c2afd989-96a7-4562-9ffd-1fac6a8f3acd","Type":"ContainerDied","Data":"d2ba9c57984ae211313cabc1cac7e28a4fc4db714de0a470aaa2cc18bf0e8792"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.109671 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-afb3-account-create-update-k5htr" event={"ID":"c2afd989-96a7-4562-9ffd-1fac6a8f3acd","Type":"ContainerStarted","Data":"1f36122cd5f4330136e893a4ec9462453438973a6d78ebed442cc396165887cf"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.115348 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" event={"ID":"f1e405ef-b09a-47b9-902e-db0ff2df7788","Type":"ContainerStarted","Data":"82433d8804f062db0a8309b9a32c6eeaa91a4c5e5b9743629079375e902c1cae"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.130542 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef75f535-8a96-40b4-9719-965a1c97a9be","Type":"ContainerStarted","Data":"de1e6e98320617a3e5b479d4b7a7572afabbab97e17ab18491f5a929d40b7f25"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.147978 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69c8df499f-wjrkv" event={"ID":"aefa115f-c019-4d47-be4c-659e19fe9eea","Type":"ContainerStarted","Data":"ed2df8acb6444d1b71421733b7082c76977cdb3527c193e5cd9910c5c9dbb63c"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.152061 4744 generic.go:334] "Generic (PLEG): container finished" podID="49796ee8-5c97-4c55-a8d8-4c47279d2641" containerID="7fc6cf3ccd66f28305831ae48e4b4a1ff31ea27fb84df4dd8b0a69e938aec37b" exitCode=0 Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.152120 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" event={"ID":"49796ee8-5c97-4c55-a8d8-4c47279d2641","Type":"ContainerDied","Data":"7fc6cf3ccd66f28305831ae48e4b4a1ff31ea27fb84df4dd8b0a69e938aec37b"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.170961 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a1b118af-7005-4a34-94ae-cf00f8c63839","Type":"ContainerDied","Data":"0536095e73f584c3907f1a4fa7e17fc39425f2fba801b6f7da26ea8a35c6bd52"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.171008 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.171026 4744 scope.go:117] "RemoveContainer" containerID="59457c62f7da9a0f199cabf8981368dc00d8eecbaa327669e025d6b51bdcb2d7" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.186946 4744 generic.go:334] "Generic (PLEG): container finished" podID="6d1c4236-2610-4bcd-91f7-c157f5da7ea4" containerID="814e34f850c30260a58a4f62d13a45bad787858a2a98db97339a39daf5eb963d" exitCode=0 Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.187039 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8jcdj" event={"ID":"6d1c4236-2610-4bcd-91f7-c157f5da7ea4","Type":"ContainerDied","Data":"814e34f850c30260a58a4f62d13a45bad787858a2a98db97339a39daf5eb963d"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.194328 4744 generic.go:334] "Generic (PLEG): container finished" podID="e57295aa-50a7-478a-805c-a682bcf0386b" containerID="2f2600111ba19818cd3e8825d0359df3c02a05a77cdbe33bd933478b8cd73829" exitCode=0 Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.194493 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b7ad-account-create-update-s4ph8" event={"ID":"e57295aa-50a7-478a-805c-a682bcf0386b","Type":"ContainerDied","Data":"2f2600111ba19818cd3e8825d0359df3c02a05a77cdbe33bd933478b8cd73829"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.200061 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerStarted","Data":"a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6"} Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.291709 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.306687 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.306746 4744 scope.go:117] "RemoveContainer" containerID="6010e9d9166c24da03fda2b4eb800cfb062467531bb0a1341308cd707efc3e30" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.319232 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 06 15:00:36 crc kubenswrapper[4744]: E0106 15:00:36.319899 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" containerName="init" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.319910 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" containerName="init" Jan 06 15:00:36 crc kubenswrapper[4744]: E0106 15:00:36.319934 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.319940 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" Jan 06 15:00:36 crc kubenswrapper[4744]: E0106 15:00:36.319952 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.319959 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" Jan 06 15:00:36 crc kubenswrapper[4744]: E0106 15:00:36.319967 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.319974 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" Jan 06 15:00:36 crc kubenswrapper[4744]: E0106 15:00:36.319996 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerName="probe" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320002 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerName="probe" Jan 06 15:00:36 crc kubenswrapper[4744]: E0106 15:00:36.320014 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerName="cinder-scheduler" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320020 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerName="cinder-scheduler" Jan 06 15:00:36 crc kubenswrapper[4744]: E0106 15:00:36.320035 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" containerName="dnsmasq-dns" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320041 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" containerName="dnsmasq-dns" Jan 06 15:00:36 crc kubenswrapper[4744]: E0106 15:00:36.320055 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320060 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320311 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320329 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" containerName="dnsmasq-dns" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320340 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320347 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerName="cinder-scheduler" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320356 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320365 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320372 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" containerName="probe" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320384 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api-log" Jan 06 15:00:36 crc kubenswrapper[4744]: E0106 15:00:36.320592 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.320600 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ddda95b-00f5-4f93-aa2c-fcf94ea2bc65" containerName="barbican-api" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.321503 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.337224 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.337887 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.465845 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.465936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.465998 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-config-data\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.466033 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8hxl\" (UniqueName: \"kubernetes.io/projected/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-kube-api-access-r8hxl\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.466064 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-scripts\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.466294 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.570031 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.570436 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.570499 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.570598 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-config-data\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.570651 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8hxl\" (UniqueName: \"kubernetes.io/projected/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-kube-api-access-r8hxl\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.570693 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-scripts\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.574815 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.578084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.579770 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.581675 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-scripts\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.582431 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-config-data\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.596666 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8hxl\" (UniqueName: \"kubernetes.io/projected/1ad5afd6-6681-4bae-a1e7-a9a7a27dae40-kube-api-access-r8hxl\") pod \"cinder-scheduler-0\" (UID: \"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40\") " pod="openstack/cinder-scheduler-0" Jan 06 15:00:36 crc kubenswrapper[4744]: I0106 15:00:36.680663 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.193246 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.219417 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jvkwj" Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.219622 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jvkwj" event={"ID":"49c698d4-2488-47f1-ad67-7506ec912b38","Type":"ContainerDied","Data":"68a6fd9ddca2398e4f52fc5730cd5948836f9b5c460687af0873f8b0fbeb8285"} Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.219664 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68a6fd9ddca2398e4f52fc5730cd5948836f9b5c460687af0873f8b0fbeb8285" Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.223375 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" event={"ID":"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a","Type":"ContainerStarted","Data":"8e0c7fe7358774ebefe9cd8b3b30df82314da210072c2215775129699a862fb4"} Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.226249 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.231036 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef75f535-8a96-40b4-9719-965a1c97a9be","Type":"ContainerStarted","Data":"03509d6d2f9b61e2b655e0c95be4e823d1d3813b0cf7134d113d00dee58bab72"} Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.231601 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.259962 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" podStartSLOduration=7.259940522 podStartE2EDuration="7.259940522s" podCreationTimestamp="2026-01-06 15:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:37.2454968 +0000 UTC m=+1433.872963118" watchObservedRunningTime="2026-01-06 15:00:37.259940522 +0000 UTC m=+1433.887406840" Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.305889 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49c698d4-2488-47f1-ad67-7506ec912b38-operator-scripts\") pod \"49c698d4-2488-47f1-ad67-7506ec912b38\" (UID: \"49c698d4-2488-47f1-ad67-7506ec912b38\") " Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.306242 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv59j\" (UniqueName: \"kubernetes.io/projected/49c698d4-2488-47f1-ad67-7506ec912b38-kube-api-access-xv59j\") pod \"49c698d4-2488-47f1-ad67-7506ec912b38\" (UID: \"49c698d4-2488-47f1-ad67-7506ec912b38\") " Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.308337 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c698d4-2488-47f1-ad67-7506ec912b38-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "49c698d4-2488-47f1-ad67-7506ec912b38" (UID: "49c698d4-2488-47f1-ad67-7506ec912b38"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.408764 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49c698d4-2488-47f1-ad67-7506ec912b38-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.468135 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 06 15:00:37 crc kubenswrapper[4744]: I0106 15:00:37.996214 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1b118af-7005-4a34-94ae-cf00f8c63839" path="/var/lib/kubelet/pods/a1b118af-7005-4a34-94ae-cf00f8c63839/volumes" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.001074 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c698d4-2488-47f1-ad67-7506ec912b38-kube-api-access-xv59j" (OuterVolumeSpecName: "kube-api-access-xv59j") pod "49c698d4-2488-47f1-ad67-7506ec912b38" (UID: "49c698d4-2488-47f1-ad67-7506ec912b38"). InnerVolumeSpecName "kube-api-access-xv59j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.039705 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv59j\" (UniqueName: \"kubernetes.io/projected/49c698d4-2488-47f1-ad67-7506ec912b38-kube-api-access-xv59j\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.285369 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.343557 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b7ad-account-create-update-s4ph8" event={"ID":"e57295aa-50a7-478a-805c-a682bcf0386b","Type":"ContainerDied","Data":"7a70955ab0f4ca2944634c7bc9f8151d51550c861344c6a5cdde656727359fd6"} Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.343598 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a70955ab0f4ca2944634c7bc9f8151d51550c861344c6a5cdde656727359fd6" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.379672 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.385369 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-sk49l" event={"ID":"1174ed04-5464-450f-850f-e77c11afee5d","Type":"ContainerDied","Data":"59e4e148b4098deb90f9714997720e30e12f688d9a6199995d482509e33b14b9"} Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.385405 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59e4e148b4098deb90f9714997720e30e12f688d9a6199995d482509e33b14b9" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.385452 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-sk49l" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.386141 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.392288 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40","Type":"ContainerStarted","Data":"661d90bed69c5d0b72202c4b1771150be24225cb013f669fb52b3e8405cd2200"} Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.411305 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-afb3-account-create-update-k5htr" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.411835 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-afb3-account-create-update-k5htr" event={"ID":"c2afd989-96a7-4562-9ffd-1fac6a8f3acd","Type":"ContainerDied","Data":"1f36122cd5f4330136e893a4ec9462453438973a6d78ebed442cc396165887cf"} Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.411866 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f36122cd5f4330136e893a4ec9462453438973a6d78ebed442cc396165887cf" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.463490 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1174ed04-5464-450f-850f-e77c11afee5d-operator-scripts\") pod \"1174ed04-5464-450f-850f-e77c11afee5d\" (UID: \"1174ed04-5464-450f-850f-e77c11afee5d\") " Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.463693 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vczd\" (UniqueName: \"kubernetes.io/projected/1174ed04-5464-450f-850f-e77c11afee5d-kube-api-access-5vczd\") pod \"1174ed04-5464-450f-850f-e77c11afee5d\" (UID: \"1174ed04-5464-450f-850f-e77c11afee5d\") " Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.465027 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1174ed04-5464-450f-850f-e77c11afee5d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1174ed04-5464-450f-850f-e77c11afee5d" (UID: "1174ed04-5464-450f-850f-e77c11afee5d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.516222 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1174ed04-5464-450f-850f-e77c11afee5d-kube-api-access-5vczd" (OuterVolumeSpecName: "kube-api-access-5vczd") pod "1174ed04-5464-450f-850f-e77c11afee5d" (UID: "1174ed04-5464-450f-850f-e77c11afee5d"). InnerVolumeSpecName "kube-api-access-5vczd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.565898 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9pk6\" (UniqueName: \"kubernetes.io/projected/e57295aa-50a7-478a-805c-a682bcf0386b-kube-api-access-s9pk6\") pod \"e57295aa-50a7-478a-805c-a682bcf0386b\" (UID: \"e57295aa-50a7-478a-805c-a682bcf0386b\") " Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.565998 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-operator-scripts\") pod \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\" (UID: \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\") " Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.566039 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57295aa-50a7-478a-805c-a682bcf0386b-operator-scripts\") pod \"e57295aa-50a7-478a-805c-a682bcf0386b\" (UID: \"e57295aa-50a7-478a-805c-a682bcf0386b\") " Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.566091 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdtx7\" (UniqueName: \"kubernetes.io/projected/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-kube-api-access-xdtx7\") pod \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\" (UID: \"c2afd989-96a7-4562-9ffd-1fac6a8f3acd\") " Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.566691 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vczd\" (UniqueName: \"kubernetes.io/projected/1174ed04-5464-450f-850f-e77c11afee5d-kube-api-access-5vczd\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.566706 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1174ed04-5464-450f-850f-e77c11afee5d-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.568713 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c2afd989-96a7-4562-9ffd-1fac6a8f3acd" (UID: "c2afd989-96a7-4562-9ffd-1fac6a8f3acd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.568953 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e57295aa-50a7-478a-805c-a682bcf0386b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e57295aa-50a7-478a-805c-a682bcf0386b" (UID: "e57295aa-50a7-478a-805c-a682bcf0386b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.586694 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-kube-api-access-xdtx7" (OuterVolumeSpecName: "kube-api-access-xdtx7") pod "c2afd989-96a7-4562-9ffd-1fac6a8f3acd" (UID: "c2afd989-96a7-4562-9ffd-1fac6a8f3acd"). InnerVolumeSpecName "kube-api-access-xdtx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.598504 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e57295aa-50a7-478a-805c-a682bcf0386b-kube-api-access-s9pk6" (OuterVolumeSpecName: "kube-api-access-s9pk6") pod "e57295aa-50a7-478a-805c-a682bcf0386b" (UID: "e57295aa-50a7-478a-805c-a682bcf0386b"). InnerVolumeSpecName "kube-api-access-s9pk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.671057 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9pk6\" (UniqueName: \"kubernetes.io/projected/e57295aa-50a7-478a-805c-a682bcf0386b-kube-api-access-s9pk6\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.671093 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.671102 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57295aa-50a7-478a-805c-a682bcf0386b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.671111 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdtx7\" (UniqueName: \"kubernetes.io/projected/c2afd989-96a7-4562-9ffd-1fac6a8f3acd-kube-api-access-xdtx7\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.707208 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-56d794f5-7hvdj"] Jan 06 15:00:38 crc kubenswrapper[4744]: E0106 15:00:38.707709 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49c698d4-2488-47f1-ad67-7506ec912b38" containerName="mariadb-database-create" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.707721 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="49c698d4-2488-47f1-ad67-7506ec912b38" containerName="mariadb-database-create" Jan 06 15:00:38 crc kubenswrapper[4744]: E0106 15:00:38.707740 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1174ed04-5464-450f-850f-e77c11afee5d" containerName="mariadb-database-create" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.707746 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1174ed04-5464-450f-850f-e77c11afee5d" containerName="mariadb-database-create" Jan 06 15:00:38 crc kubenswrapper[4744]: E0106 15:00:38.707767 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2afd989-96a7-4562-9ffd-1fac6a8f3acd" containerName="mariadb-account-create-update" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.707774 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2afd989-96a7-4562-9ffd-1fac6a8f3acd" containerName="mariadb-account-create-update" Jan 06 15:00:38 crc kubenswrapper[4744]: E0106 15:00:38.707795 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e57295aa-50a7-478a-805c-a682bcf0386b" containerName="mariadb-account-create-update" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.707801 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e57295aa-50a7-478a-805c-a682bcf0386b" containerName="mariadb-account-create-update" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.707990 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e57295aa-50a7-478a-805c-a682bcf0386b" containerName="mariadb-account-create-update" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.707999 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2afd989-96a7-4562-9ffd-1fac6a8f3acd" containerName="mariadb-account-create-update" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.708010 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="49c698d4-2488-47f1-ad67-7506ec912b38" containerName="mariadb-database-create" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.708027 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1174ed04-5464-450f-850f-e77c11afee5d" containerName="mariadb-database-create" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.708823 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.753225 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-56d794f5-7hvdj"] Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.795331 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-bc6f7454-z5n9n"] Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.796900 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.822881 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6f96c86884-j6snf"] Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.824900 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:38 crc kubenswrapper[4744]: I0106 15:00:38.840932 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-bc6f7454-z5n9n"] Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.849140 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6f96c86884-j6snf"] Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.875906 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n99c\" (UniqueName: \"kubernetes.io/projected/bd990c60-8782-4505-8e74-1b99a8b854f2-kube-api-access-8n99c\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.875962 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g2pd\" (UniqueName: \"kubernetes.io/projected/f7bf873c-7171-4ec5-a141-aeabfb5501df-kube-api-access-4g2pd\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.876033 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data-custom\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.876079 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.876113 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.876177 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data-custom\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.876195 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-combined-ca-bundle\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.876222 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-combined-ca-bundle\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987612 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data-custom\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987650 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987709 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data-custom\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987730 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-combined-ca-bundle\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987760 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-combined-ca-bundle\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n99c\" (UniqueName: \"kubernetes.io/projected/bd990c60-8782-4505-8e74-1b99a8b854f2-kube-api-access-8n99c\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987865 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g2pd\" (UniqueName: \"kubernetes.io/projected/f7bf873c-7171-4ec5-a141-aeabfb5501df-kube-api-access-4g2pd\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987928 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jsp5\" (UniqueName: \"kubernetes.io/projected/cce76110-c2be-496e-92ff-b585bde98eb8-kube-api-access-9jsp5\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.987975 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data-custom\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.988001 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-combined-ca-bundle\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:38.988045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.003620 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-combined-ca-bundle\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.006189 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data-custom\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.008396 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.008798 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-combined-ca-bundle\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.008955 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data-custom\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.009757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.023862 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n99c\" (UniqueName: \"kubernetes.io/projected/bd990c60-8782-4505-8e74-1b99a8b854f2-kube-api-access-8n99c\") pod \"heat-api-bc6f7454-z5n9n\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.029003 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g2pd\" (UniqueName: \"kubernetes.io/projected/f7bf873c-7171-4ec5-a141-aeabfb5501df-kube-api-access-4g2pd\") pod \"heat-engine-56d794f5-7hvdj\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.089570 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jsp5\" (UniqueName: \"kubernetes.io/projected/cce76110-c2be-496e-92ff-b585bde98eb8-kube-api-access-9jsp5\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.089636 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-combined-ca-bundle\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.089675 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.089706 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data-custom\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.103607 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-combined-ca-bundle\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.107943 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data-custom\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.124422 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.132928 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jsp5\" (UniqueName: \"kubernetes.io/projected/cce76110-c2be-496e-92ff-b585bde98eb8-kube-api-access-9jsp5\") pod \"heat-cfnapi-6f96c86884-j6snf\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.226443 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.243634 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.269003 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.418712 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef75f535-8a96-40b4-9719-965a1c97a9be","Type":"ContainerStarted","Data":"7ef541a63cd9e11a175782b2d049b0cd0384aed47d36b922e478c05afd4dd57c"} Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.418798 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b7ad-account-create-update-s4ph8" Jan 06 15:00:39 crc kubenswrapper[4744]: I0106 15:00:39.457968 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.457949972 podStartE2EDuration="8.457949972s" podCreationTimestamp="2026-01-06 15:00:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:39.44085965 +0000 UTC m=+1436.068325968" watchObservedRunningTime="2026-01-06 15:00:39.457949972 +0000 UTC m=+1436.085416290" Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.751486 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.773278 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.937735 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5n6z\" (UniqueName: \"kubernetes.io/projected/49796ee8-5c97-4c55-a8d8-4c47279d2641-kube-api-access-j5n6z\") pod \"49796ee8-5c97-4c55-a8d8-4c47279d2641\" (UID: \"49796ee8-5c97-4c55-a8d8-4c47279d2641\") " Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.938320 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-operator-scripts\") pod \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\" (UID: \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\") " Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.938563 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v677j\" (UniqueName: \"kubernetes.io/projected/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-kube-api-access-v677j\") pod \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\" (UID: \"6d1c4236-2610-4bcd-91f7-c157f5da7ea4\") " Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.938605 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49796ee8-5c97-4c55-a8d8-4c47279d2641-operator-scripts\") pod \"49796ee8-5c97-4c55-a8d8-4c47279d2641\" (UID: \"49796ee8-5c97-4c55-a8d8-4c47279d2641\") " Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.938991 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6d1c4236-2610-4bcd-91f7-c157f5da7ea4" (UID: "6d1c4236-2610-4bcd-91f7-c157f5da7ea4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.939336 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.939777 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49796ee8-5c97-4c55-a8d8-4c47279d2641-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "49796ee8-5c97-4c55-a8d8-4c47279d2641" (UID: "49796ee8-5c97-4c55-a8d8-4c47279d2641"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.944975 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-kube-api-access-v677j" (OuterVolumeSpecName: "kube-api-access-v677j") pod "6d1c4236-2610-4bcd-91f7-c157f5da7ea4" (UID: "6d1c4236-2610-4bcd-91f7-c157f5da7ea4"). InnerVolumeSpecName "kube-api-access-v677j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:40 crc kubenswrapper[4744]: I0106 15:00:40.960152 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49796ee8-5c97-4c55-a8d8-4c47279d2641-kube-api-access-j5n6z" (OuterVolumeSpecName: "kube-api-access-j5n6z") pod "49796ee8-5c97-4c55-a8d8-4c47279d2641" (UID: "49796ee8-5c97-4c55-a8d8-4c47279d2641"). InnerVolumeSpecName "kube-api-access-j5n6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.042001 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v677j\" (UniqueName: \"kubernetes.io/projected/6d1c4236-2610-4bcd-91f7-c157f5da7ea4-kube-api-access-v677j\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.042035 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/49796ee8-5c97-4c55-a8d8-4c47279d2641-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.042047 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5n6z\" (UniqueName: \"kubernetes.io/projected/49796ee8-5c97-4c55-a8d8-4c47279d2641-kube-api-access-j5n6z\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.278125 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69c8df499f-wjrkv"] Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.307635 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-64557f6b9b-kxvsw"] Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.324183 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5985d564f7-8vfg7"] Jan 06 15:00:41 crc kubenswrapper[4744]: E0106 15:00:41.324779 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1c4236-2610-4bcd-91f7-c157f5da7ea4" containerName="mariadb-database-create" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.324798 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1c4236-2610-4bcd-91f7-c157f5da7ea4" containerName="mariadb-database-create" Jan 06 15:00:41 crc kubenswrapper[4744]: E0106 15:00:41.324812 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49796ee8-5c97-4c55-a8d8-4c47279d2641" containerName="mariadb-account-create-update" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.324819 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="49796ee8-5c97-4c55-a8d8-4c47279d2641" containerName="mariadb-account-create-update" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.325031 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1c4236-2610-4bcd-91f7-c157f5da7ea4" containerName="mariadb-database-create" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.325054 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="49796ee8-5c97-4c55-a8d8-4c47279d2641" containerName="mariadb-account-create-update" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.325891 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.328842 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.329185 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.339715 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5985d564f7-8vfg7"] Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.362382 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.388865 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7f5fdd5db8-d2k9t"] Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.390428 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.394928 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.395605 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.437278 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7f5fdd5db8-d2k9t"] Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.449559 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-public-tls-certs\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.449623 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjd2g\" (UniqueName: \"kubernetes.io/projected/83192382-3f34-4716-b10b-123923752ba5-kube-api-access-pjd2g\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.449678 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.449812 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-internal-tls-certs\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.449837 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data-custom\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.449869 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-combined-ca-bundle\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.456252 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40","Type":"ContainerStarted","Data":"22ebe45553e74bab07d97869f47db77fac371e55ecaf71eabd242c66996f9e38"} Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.461583 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8jcdj" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.461688 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8jcdj" event={"ID":"6d1c4236-2610-4bcd-91f7-c157f5da7ea4","Type":"ContainerDied","Data":"d6568f22b1f89e404ca83ba13ff0b35f5885680e2287d4c4739ac10a55f6da8e"} Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.461712 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6568f22b1f89e404ca83ba13ff0b35f5885680e2287d4c4739ac10a55f6da8e" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.467353 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" event={"ID":"49796ee8-5c97-4c55-a8d8-4c47279d2641","Type":"ContainerDied","Data":"3c99a231b9c99589b6a22f64c8560401e0f398548f37754aa102a4c2b373722b"} Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.467392 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c99a231b9c99589b6a22f64c8560401e0f398548f37754aa102a4c2b373722b" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.467452 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-63fc-account-create-update-bjlhb" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.479531 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-znjwg"] Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.479743 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" podUID="bf6d608b-355b-471e-9fab-40b62a25b8f0" containerName="dnsmasq-dns" containerID="cri-o://2ab1b4a19c367c9ffceebc495e6f631a1fd7c02935d8396230ea56f4d9c7162f" gracePeriod=10 Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.554792 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-combined-ca-bundle\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.554837 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-public-tls-certs\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.554936 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgwqr\" (UniqueName: \"kubernetes.io/projected/4adab37a-4969-424b-87dd-5e59dc9cd756-kube-api-access-xgwqr\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.554983 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data-custom\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.555017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-public-tls-certs\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.555059 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjd2g\" (UniqueName: \"kubernetes.io/projected/83192382-3f34-4716-b10b-123923752ba5-kube-api-access-pjd2g\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.555114 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.555207 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-internal-tls-certs\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.555307 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-internal-tls-certs\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.555336 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.555366 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data-custom\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.555404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-combined-ca-bundle\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.560271 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-internal-tls-certs\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.561756 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.562391 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-combined-ca-bundle\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.565012 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data-custom\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.575459 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-public-tls-certs\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.576012 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjd2g\" (UniqueName: \"kubernetes.io/projected/83192382-3f34-4716-b10b-123923752ba5-kube-api-access-pjd2g\") pod \"heat-api-5985d564f7-8vfg7\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.673264 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.678261 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-internal-tls-certs\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.678387 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.678475 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-combined-ca-bundle\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.678494 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-public-tls-certs\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.678542 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgwqr\" (UniqueName: \"kubernetes.io/projected/4adab37a-4969-424b-87dd-5e59dc9cd756-kube-api-access-xgwqr\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.678575 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data-custom\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.693117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-combined-ca-bundle\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.701975 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-public-tls-certs\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.704035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-internal-tls-certs\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.706219 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.715911 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgwqr\" (UniqueName: \"kubernetes.io/projected/4adab37a-4969-424b-87dd-5e59dc9cd756-kube-api-access-xgwqr\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:41 crc kubenswrapper[4744]: I0106 15:00:41.721007 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data-custom\") pod \"heat-cfnapi-7f5fdd5db8-d2k9t\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.012944 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.181779 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.181884 4744 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.237733 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.275904 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.275958 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.311439 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.329106 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.479454 4744 generic.go:334] "Generic (PLEG): container finished" podID="bf6d608b-355b-471e-9fab-40b62a25b8f0" containerID="2ab1b4a19c367c9ffceebc495e6f631a1fd7c02935d8396230ea56f4d9c7162f" exitCode=0 Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.479853 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" event={"ID":"bf6d608b-355b-471e-9fab-40b62a25b8f0","Type":"ContainerDied","Data":"2ab1b4a19c367c9ffceebc495e6f631a1fd7c02935d8396230ea56f4d9c7162f"} Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.480013 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 06 15:00:42 crc kubenswrapper[4744]: I0106 15:00:42.480039 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.513520 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" event={"ID":"bf6d608b-355b-471e-9fab-40b62a25b8f0","Type":"ContainerDied","Data":"6fb418c65bf1ed1af72f36bd8c4cb20d78214bcc63a2c286c95b3c5da6c36dc5"} Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.514058 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fb418c65bf1ed1af72f36bd8c4cb20d78214bcc63a2c286c95b3c5da6c36dc5" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.696047 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.738857 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-svc\") pod \"bf6d608b-355b-471e-9fab-40b62a25b8f0\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.738921 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-nb\") pod \"bf6d608b-355b-471e-9fab-40b62a25b8f0\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.739060 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5cl7\" (UniqueName: \"kubernetes.io/projected/bf6d608b-355b-471e-9fab-40b62a25b8f0-kube-api-access-k5cl7\") pod \"bf6d608b-355b-471e-9fab-40b62a25b8f0\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.739093 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-config\") pod \"bf6d608b-355b-471e-9fab-40b62a25b8f0\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.739250 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-swift-storage-0\") pod \"bf6d608b-355b-471e-9fab-40b62a25b8f0\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.739283 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-sb\") pod \"bf6d608b-355b-471e-9fab-40b62a25b8f0\" (UID: \"bf6d608b-355b-471e-9fab-40b62a25b8f0\") " Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.785379 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf6d608b-355b-471e-9fab-40b62a25b8f0-kube-api-access-k5cl7" (OuterVolumeSpecName: "kube-api-access-k5cl7") pod "bf6d608b-355b-471e-9fab-40b62a25b8f0" (UID: "bf6d608b-355b-471e-9fab-40b62a25b8f0"). InnerVolumeSpecName "kube-api-access-k5cl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.841508 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5cl7\" (UniqueName: \"kubernetes.io/projected/bf6d608b-355b-471e-9fab-40b62a25b8f0-kube-api-access-k5cl7\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.924472 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-config" (OuterVolumeSpecName: "config") pod "bf6d608b-355b-471e-9fab-40b62a25b8f0" (UID: "bf6d608b-355b-471e-9fab-40b62a25b8f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.939368 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bf6d608b-355b-471e-9fab-40b62a25b8f0" (UID: "bf6d608b-355b-471e-9fab-40b62a25b8f0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.971016 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.971292 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.986871 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bf6d608b-355b-471e-9fab-40b62a25b8f0" (UID: "bf6d608b-355b-471e-9fab-40b62a25b8f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.987062 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bf6d608b-355b-471e-9fab-40b62a25b8f0" (UID: "bf6d608b-355b-471e-9fab-40b62a25b8f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:43 crc kubenswrapper[4744]: I0106 15:00:43.988858 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bf6d608b-355b-471e-9fab-40b62a25b8f0" (UID: "bf6d608b-355b-471e-9fab-40b62a25b8f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.075057 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.075095 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.075105 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf6d608b-355b-471e-9fab-40b62a25b8f0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.235635 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6f96c86884-j6snf"] Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.423722 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.423996 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.532266 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.541736 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-znjwg" Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.543064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f96c86884-j6snf" event={"ID":"cce76110-c2be-496e-92ff-b585bde98eb8","Type":"ContainerStarted","Data":"ac3f5244aed3dee21656335f51e76d8dae8302b455cd4d0d22134f9d9bc7efd6"} Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.634767 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-znjwg"] Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.643521 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-znjwg"] Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.874102 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-bc6f7454-z5n9n"] Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.945122 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-56d794f5-7hvdj"] Jan 06 15:00:44 crc kubenswrapper[4744]: I0106 15:00:44.991481 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5985d564f7-8vfg7"] Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.155398 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7f5fdd5db8-d2k9t"] Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.565001 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" event={"ID":"4adab37a-4969-424b-87dd-5e59dc9cd756","Type":"ContainerStarted","Data":"caccc07ce56e2fcd8e093b6cf5655bfee83402876f1534a810c1d548a1f6f1e0"} Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.574147 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerStarted","Data":"c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2"} Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.574338 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="ceilometer-central-agent" containerID="cri-o://bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f" gracePeriod=30 Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.574569 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.574904 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="proxy-httpd" containerID="cri-o://c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2" gracePeriod=30 Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.574963 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="sg-core" containerID="cri-o://a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6" gracePeriod=30 Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.575001 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="ceilometer-notification-agent" containerID="cri-o://67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9" gracePeriod=30 Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.583136 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69c8df499f-wjrkv" event={"ID":"aefa115f-c019-4d47-be4c-659e19fe9eea","Type":"ContainerStarted","Data":"3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a"} Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.583283 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-69c8df499f-wjrkv" podUID="aefa115f-c019-4d47-be4c-659e19fe9eea" containerName="heat-api" containerID="cri-o://3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a" gracePeriod=60 Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.583355 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.601526 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.183731029 podStartE2EDuration="27.601508445s" podCreationTimestamp="2026-01-06 15:00:18 +0000 UTC" firstStartedPulling="2026-01-06 15:00:19.996597078 +0000 UTC m=+1416.624063396" lastFinishedPulling="2026-01-06 15:00:43.414374494 +0000 UTC m=+1440.041840812" observedRunningTime="2026-01-06 15:00:45.600985211 +0000 UTC m=+1442.228451539" watchObservedRunningTime="2026-01-06 15:00:45.601508445 +0000 UTC m=+1442.228974763" Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.605088 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bc6f7454-z5n9n" event={"ID":"bd990c60-8782-4505-8e74-1b99a8b854f2","Type":"ContainerStarted","Data":"305ac2710be228bb6cbc8afbc58b3363f2919bbbb1081a2f7409ccb9e4412142"} Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.606636 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-56d794f5-7hvdj" event={"ID":"f7bf873c-7171-4ec5-a141-aeabfb5501df","Type":"ContainerStarted","Data":"2d592034d00e95bc1ab7277831a57867116d1ff7bf6d8587949585a287b469e5"} Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.631818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5985d564f7-8vfg7" event={"ID":"83192382-3f34-4716-b10b-123923752ba5","Type":"ContainerStarted","Data":"c74309585cbe0903bbf74b22adecd118b02581ea824df10f453720a4601bfb90"} Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.638556 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-69c8df499f-wjrkv" podStartSLOduration=7.402323351 podStartE2EDuration="15.638537075s" podCreationTimestamp="2026-01-06 15:00:30 +0000 UTC" firstStartedPulling="2026-01-06 15:00:35.185132305 +0000 UTC m=+1431.812598623" lastFinishedPulling="2026-01-06 15:00:43.421346019 +0000 UTC m=+1440.048812347" observedRunningTime="2026-01-06 15:00:45.627960235 +0000 UTC m=+1442.255426543" watchObservedRunningTime="2026-01-06 15:00:45.638537075 +0000 UTC m=+1442.266003393" Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.729515 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf6d608b-355b-471e-9fab-40b62a25b8f0" path="/var/lib/kubelet/pods/bf6d608b-355b-471e-9fab-40b62a25b8f0/volumes" Jan 06 15:00:45 crc kubenswrapper[4744]: I0106 15:00:45.798443 4744 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod80cca3b3-144e-4276-80dc-4c0ecb5c34f5"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod80cca3b3-144e-4276-80dc-4c0ecb5c34f5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod80cca3b3_144e_4276_80dc_4c0ecb5c34f5.slice" Jan 06 15:00:45 crc kubenswrapper[4744]: E0106 15:00:45.798498 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod80cca3b3-144e-4276-80dc-4c0ecb5c34f5] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod80cca3b3-144e-4276-80dc-4c0ecb5c34f5] : Timed out while waiting for systemd to remove kubepods-besteffort-pod80cca3b3_144e_4276_80dc_4c0ecb5c34f5.slice" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" podUID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.629322 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.209:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.642638 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d66f584d7-m48b5" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.669042 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-m48b5"] Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.679413 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d66f584d7-m48b5"] Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.844234 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9d6"] Jan 06 15:00:46 crc kubenswrapper[4744]: E0106 15:00:46.844817 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6d608b-355b-471e-9fab-40b62a25b8f0" containerName="dnsmasq-dns" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.844837 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6d608b-355b-471e-9fab-40b62a25b8f0" containerName="dnsmasq-dns" Jan 06 15:00:46 crc kubenswrapper[4744]: E0106 15:00:46.844893 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6d608b-355b-471e-9fab-40b62a25b8f0" containerName="init" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.844903 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6d608b-355b-471e-9fab-40b62a25b8f0" containerName="init" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.845138 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf6d608b-355b-471e-9fab-40b62a25b8f0" containerName="dnsmasq-dns" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.846035 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.850050 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.850251 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ckxdc" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.850429 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.875311 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9d6"] Jan 06 15:00:46 crc kubenswrapper[4744]: I0106 15:00:46.907748 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.000042 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-config-data\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.000153 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7b68\" (UniqueName: \"kubernetes.io/projected/6eb055cf-7e1b-441b-a1db-4e6479bc9152-kube-api-access-c7b68\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.000235 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-scripts\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.000316 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.102484 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-config-data\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.102628 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7b68\" (UniqueName: \"kubernetes.io/projected/6eb055cf-7e1b-441b-a1db-4e6479bc9152-kube-api-access-c7b68\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.102740 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-scripts\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.102855 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.109856 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-config-data\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.111703 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-scripts\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.111967 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.123691 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7b68\" (UniqueName: \"kubernetes.io/projected/6eb055cf-7e1b-441b-a1db-4e6479bc9152-kube-api-access-c7b68\") pod \"nova-cell0-conductor-db-sync-gb9d6\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.163838 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.656881 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" event={"ID":"f1e405ef-b09a-47b9-902e-db0ff2df7788","Type":"ContainerStarted","Data":"9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea"} Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.662520 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9d6"] Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.665574 4744 generic.go:334] "Generic (PLEG): container finished" podID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerID="c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2" exitCode=0 Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.665603 4744 generic.go:334] "Generic (PLEG): container finished" podID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerID="a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6" exitCode=2 Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.665626 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerDied","Data":"c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2"} Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.665652 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerDied","Data":"a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6"} Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.781757 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80cca3b3-144e-4276-80dc-4c0ecb5c34f5" path="/var/lib/kubelet/pods/80cca3b3-144e-4276-80dc-4c0ecb5c34f5/volumes" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.782794 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c5f6db8dc-vwrjb" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.877438 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.887655 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-dc6576bbb-qn5wg"] Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.887868 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-dc6576bbb-qn5wg" podUID="826e0141-c6b9-46bc-8450-88cef322d974" containerName="neutron-api" containerID="cri-o://4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3" gracePeriod=30 Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.888007 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-dc6576bbb-qn5wg" podUID="826e0141-c6b9-46bc-8450-88cef322d974" containerName="neutron-httpd" containerID="cri-o://311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489" gracePeriod=30 Jan 06 15:00:47 crc kubenswrapper[4744]: I0106 15:00:47.959550 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.688724 4744 generic.go:334] "Generic (PLEG): container finished" podID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerID="67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9" exitCode=0 Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.689012 4744 generic.go:334] "Generic (PLEG): container finished" podID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerID="bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f" exitCode=0 Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.688918 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerDied","Data":"67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.689084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerDied","Data":"bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.692051 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-56d794f5-7hvdj" event={"ID":"f7bf873c-7171-4ec5-a141-aeabfb5501df","Type":"ContainerStarted","Data":"d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.692389 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.696033 4744 generic.go:334] "Generic (PLEG): container finished" podID="826e0141-c6b9-46bc-8450-88cef322d974" containerID="311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489" exitCode=0 Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.696137 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc6576bbb-qn5wg" event={"ID":"826e0141-c6b9-46bc-8450-88cef322d974","Type":"ContainerDied","Data":"311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.698528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f96c86884-j6snf" event={"ID":"cce76110-c2be-496e-92ff-b585bde98eb8","Type":"ContainerStarted","Data":"49837ba44b56a275f99b5cb68926b5bfb9547afbe1b81a6ee5d41e9b64b6362e"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.699641 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.701444 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" event={"ID":"6eb055cf-7e1b-441b-a1db-4e6479bc9152","Type":"ContainerStarted","Data":"07d7405a66bac66f04c991feb98a52f6f30bce6fcaf61e1fd1689a56caedf024"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.703440 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1ad5afd6-6681-4bae-a1e7-a9a7a27dae40","Type":"ContainerStarted","Data":"784f73a32d8620a2eed92907c6b441de225d451ace2d617044ddb74df4d70ff1"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.715955 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-56d794f5-7hvdj" podStartSLOduration=10.71593153 podStartE2EDuration="10.71593153s" podCreationTimestamp="2026-01-06 15:00:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:48.707181039 +0000 UTC m=+1445.334647357" watchObservedRunningTime="2026-01-06 15:00:48.71593153 +0000 UTC m=+1445.343397848" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.717775 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" event={"ID":"4adab37a-4969-424b-87dd-5e59dc9cd756","Type":"ContainerStarted","Data":"cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.717915 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.719557 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5985d564f7-8vfg7" event={"ID":"83192382-3f34-4716-b10b-123923752ba5","Type":"ContainerStarted","Data":"7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.720084 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.724233 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" podUID="f1e405ef-b09a-47b9-902e-db0ff2df7788" containerName="heat-cfnapi" containerID="cri-o://9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea" gracePeriod=60 Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.724965 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bc6f7454-z5n9n" event={"ID":"bd990c60-8782-4505-8e74-1b99a8b854f2","Type":"ContainerStarted","Data":"92dcbb933ea1350c3a509c4e87a05745b2c1e7c2b0345127e1ace355bdd2ef51"} Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.724987 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.725600 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.734399 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=12.734383769 podStartE2EDuration="12.734383769s" podCreationTimestamp="2026-01-06 15:00:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:48.725912715 +0000 UTC m=+1445.353379033" watchObservedRunningTime="2026-01-06 15:00:48.734383769 +0000 UTC m=+1445.361850087" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.767499 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" podStartSLOduration=10.316016004 podStartE2EDuration="18.767481375s" podCreationTimestamp="2026-01-06 15:00:30 +0000 UTC" firstStartedPulling="2026-01-06 15:00:35.063948037 +0000 UTC m=+1431.691414355" lastFinishedPulling="2026-01-06 15:00:43.515413418 +0000 UTC m=+1440.142879726" observedRunningTime="2026-01-06 15:00:48.76542795 +0000 UTC m=+1445.392894268" watchObservedRunningTime="2026-01-06 15:00:48.767481375 +0000 UTC m=+1445.394947693" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.768226 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6f96c86884-j6snf" podStartSLOduration=10.768221734 podStartE2EDuration="10.768221734s" podCreationTimestamp="2026-01-06 15:00:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:48.742396191 +0000 UTC m=+1445.369862509" watchObservedRunningTime="2026-01-06 15:00:48.768221734 +0000 UTC m=+1445.395688052" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.787104 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5985d564f7-8vfg7" podStartSLOduration=7.787087684 podStartE2EDuration="7.787087684s" podCreationTimestamp="2026-01-06 15:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:48.777733496 +0000 UTC m=+1445.405199814" watchObservedRunningTime="2026-01-06 15:00:48.787087684 +0000 UTC m=+1445.414554002" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.791956 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" podStartSLOduration=7.791948722 podStartE2EDuration="7.791948722s" podCreationTimestamp="2026-01-06 15:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:48.791238144 +0000 UTC m=+1445.418704452" watchObservedRunningTime="2026-01-06 15:00:48.791948722 +0000 UTC m=+1445.419415040" Jan 06 15:00:48 crc kubenswrapper[4744]: I0106 15:00:48.833270 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-bc6f7454-z5n9n" podStartSLOduration=10.833249346 podStartE2EDuration="10.833249346s" podCreationTimestamp="2026-01-06 15:00:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:48.807381521 +0000 UTC m=+1445.434847839" watchObservedRunningTime="2026-01-06 15:00:48.833249346 +0000 UTC m=+1445.460715664" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.469924 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.649658 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-run-httpd\") pod \"c0405b86-e930-436b-94a0-14b08f60f6c1\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.649954 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2p96t\" (UniqueName: \"kubernetes.io/projected/c0405b86-e930-436b-94a0-14b08f60f6c1-kube-api-access-2p96t\") pod \"c0405b86-e930-436b-94a0-14b08f60f6c1\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.650108 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-config-data\") pod \"c0405b86-e930-436b-94a0-14b08f60f6c1\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.650209 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-log-httpd\") pod \"c0405b86-e930-436b-94a0-14b08f60f6c1\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.650740 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c0405b86-e930-436b-94a0-14b08f60f6c1" (UID: "c0405b86-e930-436b-94a0-14b08f60f6c1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.650799 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c0405b86-e930-436b-94a0-14b08f60f6c1" (UID: "c0405b86-e930-436b-94a0-14b08f60f6c1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.650821 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-combined-ca-bundle\") pod \"c0405b86-e930-436b-94a0-14b08f60f6c1\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.650976 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-scripts\") pod \"c0405b86-e930-436b-94a0-14b08f60f6c1\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.651022 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-sg-core-conf-yaml\") pod \"c0405b86-e930-436b-94a0-14b08f60f6c1\" (UID: \"c0405b86-e930-436b-94a0-14b08f60f6c1\") " Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.659884 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.659931 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c0405b86-e930-436b-94a0-14b08f60f6c1-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.768744 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.768803 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c0405b86-e930-436b-94a0-14b08f60f6c1","Type":"ContainerDied","Data":"002d4b36635e093b1fa011fd8aa57466323ff773fe56e2f8fd80bfc94a5305aa"} Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.768839 4744 scope.go:117] "RemoveContainer" containerID="c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.796231 4744 scope.go:117] "RemoveContainer" containerID="a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.825194 4744 scope.go:117] "RemoveContainer" containerID="67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.862024 4744 scope.go:117] "RemoveContainer" containerID="bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.925908 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0405b86-e930-436b-94a0-14b08f60f6c1-kube-api-access-2p96t" (OuterVolumeSpecName: "kube-api-access-2p96t") pod "c0405b86-e930-436b-94a0-14b08f60f6c1" (UID: "c0405b86-e930-436b-94a0-14b08f60f6c1"). InnerVolumeSpecName "kube-api-access-2p96t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.926004 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-scripts" (OuterVolumeSpecName: "scripts") pod "c0405b86-e930-436b-94a0-14b08f60f6c1" (UID: "c0405b86-e930-436b-94a0-14b08f60f6c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.965918 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.965950 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2p96t\" (UniqueName: \"kubernetes.io/projected/c0405b86-e930-436b-94a0-14b08f60f6c1-kube-api-access-2p96t\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.967369 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jf8pt"] Jan 06 15:00:49 crc kubenswrapper[4744]: E0106 15:00:49.967877 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="proxy-httpd" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.967892 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="proxy-httpd" Jan 06 15:00:49 crc kubenswrapper[4744]: E0106 15:00:49.967920 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="ceilometer-notification-agent" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.967926 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="ceilometer-notification-agent" Jan 06 15:00:49 crc kubenswrapper[4744]: E0106 15:00:49.967938 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="ceilometer-central-agent" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.967944 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="ceilometer-central-agent" Jan 06 15:00:49 crc kubenswrapper[4744]: E0106 15:00:49.967965 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="sg-core" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.967970 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="sg-core" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.968217 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="ceilometer-notification-agent" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.968233 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="proxy-httpd" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.968246 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="ceilometer-central-agent" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.968265 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" containerName="sg-core" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.970669 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.987176 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jf8pt"] Jan 06 15:00:49 crc kubenswrapper[4744]: I0106 15:00:49.988214 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c0405b86-e930-436b-94a0-14b08f60f6c1" (UID: "c0405b86-e930-436b-94a0-14b08f60f6c1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.063046 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0405b86-e930-436b-94a0-14b08f60f6c1" (UID: "c0405b86-e930-436b-94a0-14b08f60f6c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.068222 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.068261 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.170693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-utilities\") pod \"community-operators-jf8pt\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.171014 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-catalog-content\") pod \"community-operators-jf8pt\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.171050 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt6k6\" (UniqueName: \"kubernetes.io/projected/5405b281-4e69-400c-9702-0ba8b5a20918-kube-api-access-gt6k6\") pod \"community-operators-jf8pt\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.175190 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-config-data" (OuterVolumeSpecName: "config-data") pod "c0405b86-e930-436b-94a0-14b08f60f6c1" (UID: "c0405b86-e930-436b-94a0-14b08f60f6c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.273399 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-utilities\") pod \"community-operators-jf8pt\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.273442 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-catalog-content\") pod \"community-operators-jf8pt\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.273488 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt6k6\" (UniqueName: \"kubernetes.io/projected/5405b281-4e69-400c-9702-0ba8b5a20918-kube-api-access-gt6k6\") pod \"community-operators-jf8pt\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.273601 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0405b86-e930-436b-94a0-14b08f60f6c1-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.274054 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-utilities\") pod \"community-operators-jf8pt\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.274094 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-catalog-content\") pod \"community-operators-jf8pt\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.293326 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt6k6\" (UniqueName: \"kubernetes.io/projected/5405b281-4e69-400c-9702-0ba8b5a20918-kube-api-access-gt6k6\") pod \"community-operators-jf8pt\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.380239 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.475217 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.508866 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.614546 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.618522 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.624603 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.624792 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.638666 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.786572 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.786888 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.786966 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-config-data\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.786981 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-log-httpd\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.787028 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-scripts\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.787063 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-run-httpd\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.787119 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dbwh\" (UniqueName: \"kubernetes.io/projected/d10fd623-6853-4012-8102-649cb3f8fe93-kube-api-access-4dbwh\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.888969 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-scripts\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.889140 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-run-httpd\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.889268 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dbwh\" (UniqueName: \"kubernetes.io/projected/d10fd623-6853-4012-8102-649cb3f8fe93-kube-api-access-4dbwh\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.889364 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.889397 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.889476 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-config-data\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.889491 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-log-httpd\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.889885 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-log-httpd\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.892686 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-run-httpd\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.894952 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-scripts\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.896480 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-config-data\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.896789 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.911722 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:50 crc kubenswrapper[4744]: I0106 15:00:50.913025 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dbwh\" (UniqueName: \"kubernetes.io/projected/d10fd623-6853-4012-8102-649cb3f8fe93-kube-api-access-4dbwh\") pod \"ceilometer-0\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " pod="openstack/ceilometer-0" Jan 06 15:00:51 crc kubenswrapper[4744]: I0106 15:00:51.001707 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:00:51 crc kubenswrapper[4744]: I0106 15:00:51.271299 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:00:51 crc kubenswrapper[4744]: I0106 15:00:51.588148 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.209:8776/healthcheck\": dial tcp 10.217.0.209:8776: connect: connection refused" Jan 06 15:00:51 crc kubenswrapper[4744]: I0106 15:00:51.690860 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 06 15:00:51 crc kubenswrapper[4744]: I0106 15:00:51.693690 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="1ad5afd6-6681-4bae-a1e7-a9a7a27dae40" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.224:8080/\": dial tcp 10.217.0.224:8080: connect: connection refused" Jan 06 15:00:51 crc kubenswrapper[4744]: I0106 15:00:51.828474 4744 generic.go:334] "Generic (PLEG): container finished" podID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerID="6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9" exitCode=137 Jan 06 15:00:51 crc kubenswrapper[4744]: I0106 15:00:51.837397 4744 generic.go:334] "Generic (PLEG): container finished" podID="f1e405ef-b09a-47b9-902e-db0ff2df7788" containerID="9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea" exitCode=0 Jan 06 15:00:52 crc kubenswrapper[4744]: E0106 15:00:52.037592 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.058825 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0405b86-e930-436b-94a0-14b08f60f6c1" path="/var/lib/kubelet/pods/c0405b86-e930-436b-94a0-14b08f60f6c1/volumes" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.060744 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.060771 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2c546315-1cd0-45e3-809b-c6a8d2e509b3","Type":"ContainerDied","Data":"6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9"} Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.060791 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" event={"ID":"f1e405ef-b09a-47b9-902e-db0ff2df7788","Type":"ContainerDied","Data":"9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea"} Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.470782 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jf8pt"] Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.706336 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.718508 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c546315-1cd0-45e3-809b-c6a8d2e509b3-logs\") pod \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.718601 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data\") pod \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.718701 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqdrl\" (UniqueName: \"kubernetes.io/projected/2c546315-1cd0-45e3-809b-c6a8d2e509b3-kube-api-access-pqdrl\") pod \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.718771 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c546315-1cd0-45e3-809b-c6a8d2e509b3-etc-machine-id\") pod \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.719075 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-scripts\") pod \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.719116 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-combined-ca-bundle\") pod \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.719166 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data-custom\") pod \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\" (UID: \"2c546315-1cd0-45e3-809b-c6a8d2e509b3\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.719613 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2c546315-1cd0-45e3-809b-c6a8d2e509b3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2c546315-1cd0-45e3-809b-c6a8d2e509b3" (UID: "2c546315-1cd0-45e3-809b-c6a8d2e509b3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.719779 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.720011 4744 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2c546315-1cd0-45e3-809b-c6a8d2e509b3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.720090 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c546315-1cd0-45e3-809b-c6a8d2e509b3-logs" (OuterVolumeSpecName: "logs") pod "2c546315-1cd0-45e3-809b-c6a8d2e509b3" (UID: "2c546315-1cd0-45e3-809b-c6a8d2e509b3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.731378 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c546315-1cd0-45e3-809b-c6a8d2e509b3-kube-api-access-pqdrl" (OuterVolumeSpecName: "kube-api-access-pqdrl") pod "2c546315-1cd0-45e3-809b-c6a8d2e509b3" (UID: "2c546315-1cd0-45e3-809b-c6a8d2e509b3"). InnerVolumeSpecName "kube-api-access-pqdrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.736988 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-scripts" (OuterVolumeSpecName: "scripts") pod "2c546315-1cd0-45e3-809b-c6a8d2e509b3" (UID: "2c546315-1cd0-45e3-809b-c6a8d2e509b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.757409 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2c546315-1cd0-45e3-809b-c6a8d2e509b3" (UID: "2c546315-1cd0-45e3-809b-c6a8d2e509b3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.808753 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c546315-1cd0-45e3-809b-c6a8d2e509b3" (UID: "2c546315-1cd0-45e3-809b-c6a8d2e509b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.824837 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-combined-ca-bundle\") pod \"f1e405ef-b09a-47b9-902e-db0ff2df7788\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.825027 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb26c\" (UniqueName: \"kubernetes.io/projected/f1e405ef-b09a-47b9-902e-db0ff2df7788-kube-api-access-cb26c\") pod \"f1e405ef-b09a-47b9-902e-db0ff2df7788\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.825238 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data\") pod \"f1e405ef-b09a-47b9-902e-db0ff2df7788\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.825448 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data-custom\") pod \"f1e405ef-b09a-47b9-902e-db0ff2df7788\" (UID: \"f1e405ef-b09a-47b9-902e-db0ff2df7788\") " Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.826127 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.826144 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.826154 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.826176 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2c546315-1cd0-45e3-809b-c6a8d2e509b3-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.826184 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqdrl\" (UniqueName: \"kubernetes.io/projected/2c546315-1cd0-45e3-809b-c6a8d2e509b3-kube-api-access-pqdrl\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.838105 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f1e405ef-b09a-47b9-902e-db0ff2df7788" (UID: "f1e405ef-b09a-47b9-902e-db0ff2df7788"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.838415 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e405ef-b09a-47b9-902e-db0ff2df7788-kube-api-access-cb26c" (OuterVolumeSpecName: "kube-api-access-cb26c") pod "f1e405ef-b09a-47b9-902e-db0ff2df7788" (UID: "f1e405ef-b09a-47b9-902e-db0ff2df7788"). InnerVolumeSpecName "kube-api-access-cb26c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.889597 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerStarted","Data":"59b9493d3482d8b531877f984e248e6a849ce255a8f382f8304630d4f8424c55"} Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.895135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8pt" event={"ID":"5405b281-4e69-400c-9702-0ba8b5a20918","Type":"ContainerStarted","Data":"b8b01bdb832da24a924dd91c5bbf268f31e96becaaa4821ea03fd87939105af1"} Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.896542 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" event={"ID":"f1e405ef-b09a-47b9-902e-db0ff2df7788","Type":"ContainerDied","Data":"82433d8804f062db0a8309b9a32c6eeaa91a4c5e5b9743629079375e902c1cae"} Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.896580 4744 scope.go:117] "RemoveContainer" containerID="9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.896692 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-64557f6b9b-kxvsw" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.904103 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2c546315-1cd0-45e3-809b-c6a8d2e509b3","Type":"ContainerDied","Data":"7a1eabb1fd46fa9eaa70c0863168b519b6bb07624f9e45e231fc57023afbe68d"} Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.904440 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.917433 4744 generic.go:334] "Generic (PLEG): container finished" podID="bd990c60-8782-4505-8e74-1b99a8b854f2" containerID="92dcbb933ea1350c3a509c4e87a05745b2c1e7c2b0345127e1ace355bdd2ef51" exitCode=1 Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.917524 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bc6f7454-z5n9n" event={"ID":"bd990c60-8782-4505-8e74-1b99a8b854f2","Type":"ContainerDied","Data":"92dcbb933ea1350c3a509c4e87a05745b2c1e7c2b0345127e1ace355bdd2ef51"} Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.918342 4744 scope.go:117] "RemoveContainer" containerID="92dcbb933ea1350c3a509c4e87a05745b2c1e7c2b0345127e1ace355bdd2ef51" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.921881 4744 generic.go:334] "Generic (PLEG): container finished" podID="826e0141-c6b9-46bc-8450-88cef322d974" containerID="4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3" exitCode=0 Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.921940 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc6576bbb-qn5wg" event={"ID":"826e0141-c6b9-46bc-8450-88cef322d974","Type":"ContainerDied","Data":"4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3"} Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.923352 4744 generic.go:334] "Generic (PLEG): container finished" podID="cce76110-c2be-496e-92ff-b585bde98eb8" containerID="49837ba44b56a275f99b5cb68926b5bfb9547afbe1b81a6ee5d41e9b64b6362e" exitCode=1 Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.923379 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f96c86884-j6snf" event={"ID":"cce76110-c2be-496e-92ff-b585bde98eb8","Type":"ContainerDied","Data":"49837ba44b56a275f99b5cb68926b5bfb9547afbe1b81a6ee5d41e9b64b6362e"} Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.924282 4744 scope.go:117] "RemoveContainer" containerID="49837ba44b56a275f99b5cb68926b5bfb9547afbe1b81a6ee5d41e9b64b6362e" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.929444 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.929472 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb26c\" (UniqueName: \"kubernetes.io/projected/f1e405ef-b09a-47b9-902e-db0ff2df7788-kube-api-access-cb26c\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:52 crc kubenswrapper[4744]: I0106 15:00:52.981443 4744 scope.go:117] "RemoveContainer" containerID="6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.051054 4744 scope.go:117] "RemoveContainer" containerID="893d538ffec4423a138a93f8dc4f69cc32e6080f5b2602cde0634ce61a3cd4d1" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.116371 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1e405ef-b09a-47b9-902e-db0ff2df7788" (UID: "f1e405ef-b09a-47b9-902e-db0ff2df7788"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.126938 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data" (OuterVolumeSpecName: "config-data") pod "2c546315-1cd0-45e3-809b-c6a8d2e509b3" (UID: "2c546315-1cd0-45e3-809b-c6a8d2e509b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.138914 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.138943 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c546315-1cd0-45e3-809b-c6a8d2e509b3-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.185596 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data" (OuterVolumeSpecName: "config-data") pod "f1e405ef-b09a-47b9-902e-db0ff2df7788" (UID: "f1e405ef-b09a-47b9-902e-db0ff2df7788"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.241768 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e405ef-b09a-47b9-902e-db0ff2df7788-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.316941 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-64557f6b9b-kxvsw"] Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.335465 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-64557f6b9b-kxvsw"] Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.369275 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.398516 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.406644 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 06 15:00:53 crc kubenswrapper[4744]: E0106 15:00:53.407140 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api-log" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.407152 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api-log" Jan 06 15:00:53 crc kubenswrapper[4744]: E0106 15:00:53.407188 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e405ef-b09a-47b9-902e-db0ff2df7788" containerName="heat-cfnapi" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.407195 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e405ef-b09a-47b9-902e-db0ff2df7788" containerName="heat-cfnapi" Jan 06 15:00:53 crc kubenswrapper[4744]: E0106 15:00:53.407227 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.407241 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.407433 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api-log" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.407461 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" containerName="cinder-api" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.407478 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e405ef-b09a-47b9-902e-db0ff2df7788" containerName="heat-cfnapi" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.408711 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.414681 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.414901 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.414909 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.442466 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.450368 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.450421 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-scripts\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.450443 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-logs\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.450468 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-config-data\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.450540 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.450564 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.450617 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.450645 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.450716 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv5x6\" (UniqueName: \"kubernetes.io/projected/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-kube-api-access-vv5x6\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.552697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-config-data\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.552800 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.552833 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.552881 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.552914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.552976 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv5x6\" (UniqueName: \"kubernetes.io/projected/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-kube-api-access-vv5x6\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.553013 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.553035 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-scripts\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.553067 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-logs\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.553506 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-logs\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.554633 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.562411 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.562799 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-scripts\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.563759 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-config-data-custom\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.565035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-config-data\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.570665 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.572900 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.575183 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv5x6\" (UniqueName: \"kubernetes.io/projected/b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b-kube-api-access-vv5x6\") pod \"cinder-api-0\" (UID: \"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b\") " pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.734059 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c546315-1cd0-45e3-809b-c6a8d2e509b3" path="/var/lib/kubelet/pods/2c546315-1cd0-45e3-809b-c6a8d2e509b3/volumes" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.734857 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e405ef-b09a-47b9-902e-db0ff2df7788" path="/var/lib/kubelet/pods/f1e405ef-b09a-47b9-902e-db0ff2df7788/volumes" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.740638 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.940470 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bc6f7454-z5n9n" event={"ID":"bd990c60-8782-4505-8e74-1b99a8b854f2","Type":"ContainerStarted","Data":"03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603"} Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.940898 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.952376 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f96c86884-j6snf" event={"ID":"cce76110-c2be-496e-92ff-b585bde98eb8","Type":"ContainerStarted","Data":"4934cb90dabe164007c541855cc13b4c1e084cb7d66a8f9ed70cee5a95f93e41"} Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.953393 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.960540 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerStarted","Data":"f99ed76641c6b5de4c93710c95b2260d76b028d6b87f58fff59251657ee737fb"} Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.966777 4744 generic.go:334] "Generic (PLEG): container finished" podID="5405b281-4e69-400c-9702-0ba8b5a20918" containerID="84990ca01c1f1c6bf56e04da9e76ed95e4c33723a462ee9c5cc77e6dec6a2d0b" exitCode=0 Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.968096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8pt" event={"ID":"5405b281-4e69-400c-9702-0ba8b5a20918","Type":"ContainerDied","Data":"84990ca01c1f1c6bf56e04da9e76ed95e4c33723a462ee9c5cc77e6dec6a2d0b"} Jan 06 15:00:53 crc kubenswrapper[4744]: I0106 15:00:53.968223 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.070427 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-ovndb-tls-certs\") pod \"826e0141-c6b9-46bc-8450-88cef322d974\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.070487 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-config\") pod \"826e0141-c6b9-46bc-8450-88cef322d974\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.070628 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-combined-ca-bundle\") pod \"826e0141-c6b9-46bc-8450-88cef322d974\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.070782 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4jgd\" (UniqueName: \"kubernetes.io/projected/826e0141-c6b9-46bc-8450-88cef322d974-kube-api-access-m4jgd\") pod \"826e0141-c6b9-46bc-8450-88cef322d974\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.070817 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-httpd-config\") pod \"826e0141-c6b9-46bc-8450-88cef322d974\" (UID: \"826e0141-c6b9-46bc-8450-88cef322d974\") " Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.092474 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.127326 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "826e0141-c6b9-46bc-8450-88cef322d974" (UID: "826e0141-c6b9-46bc-8450-88cef322d974"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.161632 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/826e0141-c6b9-46bc-8450-88cef322d974-kube-api-access-m4jgd" (OuterVolumeSpecName: "kube-api-access-m4jgd") pod "826e0141-c6b9-46bc-8450-88cef322d974" (UID: "826e0141-c6b9-46bc-8450-88cef322d974"). InnerVolumeSpecName "kube-api-access-m4jgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.195839 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4jgd\" (UniqueName: \"kubernetes.io/projected/826e0141-c6b9-46bc-8450-88cef322d974-kube-api-access-m4jgd\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.195875 4744 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.244269 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "826e0141-c6b9-46bc-8450-88cef322d974" (UID: "826e0141-c6b9-46bc-8450-88cef322d974"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.299062 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.308262 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-config" (OuterVolumeSpecName: "config") pod "826e0141-c6b9-46bc-8450-88cef322d974" (UID: "826e0141-c6b9-46bc-8450-88cef322d974"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.325252 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "826e0141-c6b9-46bc-8450-88cef322d974" (UID: "826e0141-c6b9-46bc-8450-88cef322d974"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.400959 4744 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.400992 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/826e0141-c6b9-46bc-8450-88cef322d974-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.574368 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.605634 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.687584 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6f96c86884-j6snf"] Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.996937 4744 generic.go:334] "Generic (PLEG): container finished" podID="bd990c60-8782-4505-8e74-1b99a8b854f2" containerID="03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603" exitCode=1 Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.996996 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bc6f7454-z5n9n" event={"ID":"bd990c60-8782-4505-8e74-1b99a8b854f2","Type":"ContainerDied","Data":"03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603"} Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.997033 4744 scope.go:117] "RemoveContainer" containerID="92dcbb933ea1350c3a509c4e87a05745b2c1e7c2b0345127e1ace355bdd2ef51" Jan 06 15:00:54 crc kubenswrapper[4744]: I0106 15:00:54.997869 4744 scope.go:117] "RemoveContainer" containerID="03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603" Jan 06 15:00:54 crc kubenswrapper[4744]: E0106 15:00:54.998092 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-bc6f7454-z5n9n_openstack(bd990c60-8782-4505-8e74-1b99a8b854f2)\"" pod="openstack/heat-api-bc6f7454-z5n9n" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.018479 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b","Type":"ContainerStarted","Data":"b7b527392574557ddb3a5deb684187c6a0e76ac70d1e8a7c2869611a342fcb24"} Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.026908 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dc6576bbb-qn5wg" event={"ID":"826e0141-c6b9-46bc-8450-88cef322d974","Type":"ContainerDied","Data":"2b06a91900b437fe3ccafd75d3bf8483ade95942fcff3763072c6fbcdd086cf4"} Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.027019 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dc6576bbb-qn5wg" Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.032719 4744 generic.go:334] "Generic (PLEG): container finished" podID="cce76110-c2be-496e-92ff-b585bde98eb8" containerID="4934cb90dabe164007c541855cc13b4c1e084cb7d66a8f9ed70cee5a95f93e41" exitCode=1 Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.032758 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f96c86884-j6snf" event={"ID":"cce76110-c2be-496e-92ff-b585bde98eb8","Type":"ContainerDied","Data":"4934cb90dabe164007c541855cc13b4c1e084cb7d66a8f9ed70cee5a95f93e41"} Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.033309 4744 scope.go:117] "RemoveContainer" containerID="4934cb90dabe164007c541855cc13b4c1e084cb7d66a8f9ed70cee5a95f93e41" Jan 06 15:00:55 crc kubenswrapper[4744]: E0106 15:00:55.033596 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6f96c86884-j6snf_openstack(cce76110-c2be-496e-92ff-b585bde98eb8)\"" pod="openstack/heat-cfnapi-6f96c86884-j6snf" podUID="cce76110-c2be-496e-92ff-b585bde98eb8" Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.109664 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-dc6576bbb-qn5wg"] Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.119509 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-dc6576bbb-qn5wg"] Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.656561 4744 scope.go:117] "RemoveContainer" containerID="311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489" Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.727631 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="826e0141-c6b9-46bc-8450-88cef322d974" path="/var/lib/kubelet/pods/826e0141-c6b9-46bc-8450-88cef322d974/volumes" Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.751183 4744 scope.go:117] "RemoveContainer" containerID="4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3" Jan 06 15:00:55 crc kubenswrapper[4744]: I0106 15:00:55.815041 4744 scope.go:117] "RemoveContainer" containerID="49837ba44b56a275f99b5cb68926b5bfb9547afbe1b81a6ee5d41e9b64b6362e" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.073089 4744 scope.go:117] "RemoveContainer" containerID="03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603" Jan 06 15:00:56 crc kubenswrapper[4744]: E0106 15:00:56.073672 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-bc6f7454-z5n9n_openstack(bd990c60-8782-4505-8e74-1b99a8b854f2)\"" pod="openstack/heat-api-bc6f7454-z5n9n" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.113639 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b","Type":"ContainerStarted","Data":"d486c14500b25461b0dda6ae3180f963366a1e68c210aa0edcb58312a937deca"} Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.687576 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.764774 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-combined-ca-bundle\") pod \"cce76110-c2be-496e-92ff-b585bde98eb8\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.764866 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data\") pod \"cce76110-c2be-496e-92ff-b585bde98eb8\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.764990 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data-custom\") pod \"cce76110-c2be-496e-92ff-b585bde98eb8\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.765042 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jsp5\" (UniqueName: \"kubernetes.io/projected/cce76110-c2be-496e-92ff-b585bde98eb8-kube-api-access-9jsp5\") pod \"cce76110-c2be-496e-92ff-b585bde98eb8\" (UID: \"cce76110-c2be-496e-92ff-b585bde98eb8\") " Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.770473 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cce76110-c2be-496e-92ff-b585bde98eb8-kube-api-access-9jsp5" (OuterVolumeSpecName: "kube-api-access-9jsp5") pod "cce76110-c2be-496e-92ff-b585bde98eb8" (UID: "cce76110-c2be-496e-92ff-b585bde98eb8"). InnerVolumeSpecName "kube-api-access-9jsp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.774901 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cce76110-c2be-496e-92ff-b585bde98eb8" (UID: "cce76110-c2be-496e-92ff-b585bde98eb8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.868298 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.868329 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jsp5\" (UniqueName: \"kubernetes.io/projected/cce76110-c2be-496e-92ff-b585bde98eb8-kube-api-access-9jsp5\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.913294 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data" (OuterVolumeSpecName: "config-data") pod "cce76110-c2be-496e-92ff-b585bde98eb8" (UID: "cce76110-c2be-496e-92ff-b585bde98eb8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.921488 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cce76110-c2be-496e-92ff-b585bde98eb8" (UID: "cce76110-c2be-496e-92ff-b585bde98eb8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.970142 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:56 crc kubenswrapper[4744]: I0106 15:00:56.970196 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cce76110-c2be-496e-92ff-b585bde98eb8-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.130351 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8pt" event={"ID":"5405b281-4e69-400c-9702-0ba8b5a20918","Type":"ContainerStarted","Data":"7c16cea3773de90e3b6682b5c0afa4a55dfbc394fa88316d7d785a2557cb6a60"} Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.147355 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f96c86884-j6snf" Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.147448 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f96c86884-j6snf" event={"ID":"cce76110-c2be-496e-92ff-b585bde98eb8","Type":"ContainerDied","Data":"ac3f5244aed3dee21656335f51e76d8dae8302b455cd4d0d22134f9d9bc7efd6"} Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.147501 4744 scope.go:117] "RemoveContainer" containerID="4934cb90dabe164007c541855cc13b4c1e084cb7d66a8f9ed70cee5a95f93e41" Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.152928 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.163521 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerStarted","Data":"d365e3d5aee297c5b1288d55446a35840c4d302794f56ad081a607472297daf8"} Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.254879 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6f96c86884-j6snf"] Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.287732 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-6f96c86884-j6snf"] Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.723847 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cce76110-c2be-496e-92ff-b585bde98eb8" path="/var/lib/kubelet/pods/cce76110-c2be-496e-92ff-b585bde98eb8/volumes" Jan 06 15:00:57 crc kubenswrapper[4744]: I0106 15:00:57.853339 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:00:58 crc kubenswrapper[4744]: I0106 15:00:58.174785 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b","Type":"ContainerStarted","Data":"f287163487fa771184280aa4faba041d34c3aa61e253cf104f73232f70a72496"} Jan 06 15:00:59 crc kubenswrapper[4744]: I0106 15:00:59.187242 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 06 15:00:59 crc kubenswrapper[4744]: I0106 15:00:59.244903 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:00:59 crc kubenswrapper[4744]: I0106 15:00:59.245868 4744 scope.go:117] "RemoveContainer" containerID="03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603" Jan 06 15:00:59 crc kubenswrapper[4744]: E0106 15:00:59.246220 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-bc6f7454-z5n9n_openstack(bd990c60-8782-4505-8e74-1b99a8b854f2)\"" pod="openstack/heat-api-bc6f7454-z5n9n" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" Jan 06 15:00:59 crc kubenswrapper[4744]: I0106 15:00:59.262754 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:00:59 crc kubenswrapper[4744]: I0106 15:00:59.285739 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.285721371 podStartE2EDuration="6.285721371s" podCreationTimestamp="2026-01-06 15:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:00:59.211551848 +0000 UTC m=+1455.839018176" watchObservedRunningTime="2026-01-06 15:00:59.285721371 +0000 UTC m=+1455.913187689" Jan 06 15:00:59 crc kubenswrapper[4744]: I0106 15:00:59.318789 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7bd5fcfc47-lf5fz"] Jan 06 15:00:59 crc kubenswrapper[4744]: I0106 15:00:59.319009 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" podUID="13f41b35-92ba-4b8a-85cf-969aeccdba04" containerName="heat-engine" containerID="cri-o://a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6" gracePeriod=60 Jan 06 15:00:59 crc kubenswrapper[4744]: I0106 15:00:59.471344 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:00:59 crc kubenswrapper[4744]: I0106 15:00:59.546010 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-bc6f7454-z5n9n"] Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.143944 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29461861-vjkz8"] Jan 06 15:01:00 crc kubenswrapper[4744]: E0106 15:01:00.144665 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce76110-c2be-496e-92ff-b585bde98eb8" containerName="heat-cfnapi" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.144690 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce76110-c2be-496e-92ff-b585bde98eb8" containerName="heat-cfnapi" Jan 06 15:01:00 crc kubenswrapper[4744]: E0106 15:01:00.144704 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826e0141-c6b9-46bc-8450-88cef322d974" containerName="neutron-httpd" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.144713 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="826e0141-c6b9-46bc-8450-88cef322d974" containerName="neutron-httpd" Jan 06 15:01:00 crc kubenswrapper[4744]: E0106 15:01:00.144741 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="826e0141-c6b9-46bc-8450-88cef322d974" containerName="neutron-api" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.144750 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="826e0141-c6b9-46bc-8450-88cef322d974" containerName="neutron-api" Jan 06 15:01:00 crc kubenswrapper[4744]: E0106 15:01:00.144785 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cce76110-c2be-496e-92ff-b585bde98eb8" containerName="heat-cfnapi" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.144794 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="cce76110-c2be-496e-92ff-b585bde98eb8" containerName="heat-cfnapi" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.145124 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cce76110-c2be-496e-92ff-b585bde98eb8" containerName="heat-cfnapi" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.145149 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="cce76110-c2be-496e-92ff-b585bde98eb8" containerName="heat-cfnapi" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.145183 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="826e0141-c6b9-46bc-8450-88cef322d974" containerName="neutron-api" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.145193 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="826e0141-c6b9-46bc-8450-88cef322d974" containerName="neutron-httpd" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.146222 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.155080 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29461861-vjkz8"] Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.256939 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-fernet-keys\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.257366 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-config-data\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.257433 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-combined-ca-bundle\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.257472 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqq8g\" (UniqueName: \"kubernetes.io/projected/d3548869-d949-4d99-8968-8cdf8b347771-kube-api-access-rqq8g\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.359783 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-config-data\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.359876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-combined-ca-bundle\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.359961 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqq8g\" (UniqueName: \"kubernetes.io/projected/d3548869-d949-4d99-8968-8cdf8b347771-kube-api-access-rqq8g\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.360054 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-fernet-keys\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.367199 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-fernet-keys\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.371574 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-combined-ca-bundle\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.381652 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-config-data\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.387850 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqq8g\" (UniqueName: \"kubernetes.io/projected/d3548869-d949-4d99-8968-8cdf8b347771-kube-api-access-rqq8g\") pod \"keystone-cron-29461861-vjkz8\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.483926 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.881873 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.980136 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n99c\" (UniqueName: \"kubernetes.io/projected/bd990c60-8782-4505-8e74-1b99a8b854f2-kube-api-access-8n99c\") pod \"bd990c60-8782-4505-8e74-1b99a8b854f2\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.980469 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data\") pod \"bd990c60-8782-4505-8e74-1b99a8b854f2\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.980555 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-combined-ca-bundle\") pod \"bd990c60-8782-4505-8e74-1b99a8b854f2\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.980585 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data-custom\") pod \"bd990c60-8782-4505-8e74-1b99a8b854f2\" (UID: \"bd990c60-8782-4505-8e74-1b99a8b854f2\") " Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.986573 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bd990c60-8782-4505-8e74-1b99a8b854f2" (UID: "bd990c60-8782-4505-8e74-1b99a8b854f2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:00 crc kubenswrapper[4744]: I0106 15:01:00.986929 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd990c60-8782-4505-8e74-1b99a8b854f2-kube-api-access-8n99c" (OuterVolumeSpecName: "kube-api-access-8n99c") pod "bd990c60-8782-4505-8e74-1b99a8b854f2" (UID: "bd990c60-8782-4505-8e74-1b99a8b854f2"). InnerVolumeSpecName "kube-api-access-8n99c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.030262 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd990c60-8782-4505-8e74-1b99a8b854f2" (UID: "bd990c60-8782-4505-8e74-1b99a8b854f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:01 crc kubenswrapper[4744]: E0106 15:01:01.052926 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:01:01 crc kubenswrapper[4744]: E0106 15:01:01.056075 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.065324 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data" (OuterVolumeSpecName: "config-data") pod "bd990c60-8782-4505-8e74-1b99a8b854f2" (UID: "bd990c60-8782-4505-8e74-1b99a8b854f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:01 crc kubenswrapper[4744]: E0106 15:01:01.065415 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:01:01 crc kubenswrapper[4744]: E0106 15:01:01.065456 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" podUID="13f41b35-92ba-4b8a-85cf-969aeccdba04" containerName="heat-engine" Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.083632 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n99c\" (UniqueName: \"kubernetes.io/projected/bd990c60-8782-4505-8e74-1b99a8b854f2-kube-api-access-8n99c\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.083664 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.083674 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.083682 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bd990c60-8782-4505-8e74-1b99a8b854f2-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.236303 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-bc6f7454-z5n9n" event={"ID":"bd990c60-8782-4505-8e74-1b99a8b854f2","Type":"ContainerDied","Data":"305ac2710be228bb6cbc8afbc58b3363f2919bbbb1081a2f7409ccb9e4412142"} Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.236349 4744 scope.go:117] "RemoveContainer" containerID="03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603" Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.236444 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-bc6f7454-z5n9n" Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.243989 4744 generic.go:334] "Generic (PLEG): container finished" podID="5405b281-4e69-400c-9702-0ba8b5a20918" containerID="7c16cea3773de90e3b6682b5c0afa4a55dfbc394fa88316d7d785a2557cb6a60" exitCode=0 Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.244049 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8pt" event={"ID":"5405b281-4e69-400c-9702-0ba8b5a20918","Type":"ContainerDied","Data":"7c16cea3773de90e3b6682b5c0afa4a55dfbc394fa88316d7d785a2557cb6a60"} Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.298099 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-bc6f7454-z5n9n"] Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.308365 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-bc6f7454-z5n9n"] Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.379416 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29461861-vjkz8"] Jan 06 15:01:01 crc kubenswrapper[4744]: I0106 15:01:01.723334 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" path="/var/lib/kubelet/pods/bd990c60-8782-4505-8e74-1b99a8b854f2/volumes" Jan 06 15:01:02 crc kubenswrapper[4744]: I0106 15:01:02.258284 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29461861-vjkz8" event={"ID":"d3548869-d949-4d99-8968-8cdf8b347771","Type":"ContainerStarted","Data":"2bb0bdf38cb8dda23637d5cf1ecfd04610da8f565b428fdff7d30d368ae0e21d"} Jan 06 15:01:04 crc kubenswrapper[4744]: I0106 15:01:04.024822 4744 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pode39dd03e-1fc4-4650-b0b6-d0067a81ed77"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pode39dd03e-1fc4-4650-b0b6-d0067a81ed77] : Timed out while waiting for systemd to remove kubepods-besteffort-pode39dd03e_1fc4_4650_b0b6_d0067a81ed77.slice" Jan 06 15:01:04 crc kubenswrapper[4744]: E0106 15:01:04.026287 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pode39dd03e-1fc4-4650-b0b6-d0067a81ed77] : unable to destroy cgroup paths for cgroup [kubepods besteffort pode39dd03e-1fc4-4650-b0b6-d0067a81ed77] : Timed out while waiting for systemd to remove kubepods-besteffort-pode39dd03e_1fc4_4650_b0b6_d0067a81ed77.slice" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" podUID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" Jan 06 15:01:04 crc kubenswrapper[4744]: I0106 15:01:04.290245 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-gndpz" Jan 06 15:01:04 crc kubenswrapper[4744]: I0106 15:01:04.593861 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-gndpz"] Jan 06 15:01:04 crc kubenswrapper[4744]: I0106 15:01:04.604323 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-gndpz"] Jan 06 15:01:05 crc kubenswrapper[4744]: I0106 15:01:05.303867 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29461861-vjkz8" event={"ID":"d3548869-d949-4d99-8968-8cdf8b347771","Type":"ContainerStarted","Data":"e646352e09eaa75db2837a582edc24559bbacee33cf11c0e1c966dde108074c3"} Jan 06 15:01:05 crc kubenswrapper[4744]: I0106 15:01:05.345717 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29461861-vjkz8" podStartSLOduration=5.345693642 podStartE2EDuration="5.345693642s" podCreationTimestamp="2026-01-06 15:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:01:05.338413799 +0000 UTC m=+1461.965880117" watchObservedRunningTime="2026-01-06 15:01:05.345693642 +0000 UTC m=+1461.973159960" Jan 06 15:01:05 crc kubenswrapper[4744]: I0106 15:01:05.724338 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e39dd03e-1fc4-4650-b0b6-d0067a81ed77" path="/var/lib/kubelet/pods/e39dd03e-1fc4-4650-b0b6-d0067a81ed77/volumes" Jan 06 15:01:07 crc kubenswrapper[4744]: I0106 15:01:07.750391 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.233:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:01:08 crc kubenswrapper[4744]: I0106 15:01:08.750379 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.233:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:01:10 crc kubenswrapper[4744]: I0106 15:01:10.368709 4744 generic.go:334] "Generic (PLEG): container finished" podID="d3548869-d949-4d99-8968-8cdf8b347771" containerID="e646352e09eaa75db2837a582edc24559bbacee33cf11c0e1c966dde108074c3" exitCode=0 Jan 06 15:01:10 crc kubenswrapper[4744]: I0106 15:01:10.368786 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29461861-vjkz8" event={"ID":"d3548869-d949-4d99-8968-8cdf8b347771","Type":"ContainerDied","Data":"e646352e09eaa75db2837a582edc24559bbacee33cf11c0e1c966dde108074c3"} Jan 06 15:01:11 crc kubenswrapper[4744]: E0106 15:01:11.039376 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:01:11 crc kubenswrapper[4744]: E0106 15:01:11.043258 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:01:11 crc kubenswrapper[4744]: E0106 15:01:11.045471 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:01:11 crc kubenswrapper[4744]: E0106 15:01:11.045504 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" podUID="13f41b35-92ba-4b8a-85cf-969aeccdba04" containerName="heat-engine" Jan 06 15:01:11 crc kubenswrapper[4744]: I0106 15:01:11.291589 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.081436 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.233562 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-combined-ca-bundle\") pod \"d3548869-d949-4d99-8968-8cdf8b347771\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.233798 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-fernet-keys\") pod \"d3548869-d949-4d99-8968-8cdf8b347771\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.233973 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqq8g\" (UniqueName: \"kubernetes.io/projected/d3548869-d949-4d99-8968-8cdf8b347771-kube-api-access-rqq8g\") pod \"d3548869-d949-4d99-8968-8cdf8b347771\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.234085 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-config-data\") pod \"d3548869-d949-4d99-8968-8cdf8b347771\" (UID: \"d3548869-d949-4d99-8968-8cdf8b347771\") " Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.239545 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3548869-d949-4d99-8968-8cdf8b347771-kube-api-access-rqq8g" (OuterVolumeSpecName: "kube-api-access-rqq8g") pod "d3548869-d949-4d99-8968-8cdf8b347771" (UID: "d3548869-d949-4d99-8968-8cdf8b347771"). InnerVolumeSpecName "kube-api-access-rqq8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.242456 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d3548869-d949-4d99-8968-8cdf8b347771" (UID: "d3548869-d949-4d99-8968-8cdf8b347771"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:14 crc kubenswrapper[4744]: E0106 15:01:14.280792 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified" Jan 06 15:01:14 crc kubenswrapper[4744]: E0106 15:01:14.281064 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c7b68,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-gb9d6_openstack(6eb055cf-7e1b-441b-a1db-4e6479bc9152): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 15:01:14 crc kubenswrapper[4744]: E0106 15:01:14.282454 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" podUID="6eb055cf-7e1b-441b-a1db-4e6479bc9152" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.285848 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3548869-d949-4d99-8968-8cdf8b347771" (UID: "d3548869-d949-4d99-8968-8cdf8b347771"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.318483 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-config-data" (OuterVolumeSpecName: "config-data") pod "d3548869-d949-4d99-8968-8cdf8b347771" (UID: "d3548869-d949-4d99-8968-8cdf8b347771"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.337552 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqq8g\" (UniqueName: \"kubernetes.io/projected/d3548869-d949-4d99-8968-8cdf8b347771-kube-api-access-rqq8g\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.337875 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.338010 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.338143 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d3548869-d949-4d99-8968-8cdf8b347771-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.421625 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29461861-vjkz8" event={"ID":"d3548869-d949-4d99-8968-8cdf8b347771","Type":"ContainerDied","Data":"2bb0bdf38cb8dda23637d5cf1ecfd04610da8f565b428fdff7d30d368ae0e21d"} Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.421668 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bb0bdf38cb8dda23637d5cf1ecfd04610da8f565b428fdff7d30d368ae0e21d" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.421728 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29461861-vjkz8" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.429795 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.429866 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.448289 4744 generic.go:334] "Generic (PLEG): container finished" podID="13f41b35-92ba-4b8a-85cf-969aeccdba04" containerID="a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6" exitCode=0 Jan 06 15:01:14 crc kubenswrapper[4744]: I0106 15:01:14.448529 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" event={"ID":"13f41b35-92ba-4b8a-85cf-969aeccdba04","Type":"ContainerDied","Data":"a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6"} Jan 06 15:01:15 crc kubenswrapper[4744]: E0106 15:01:15.592336 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" podUID="6eb055cf-7e1b-441b-a1db-4e6479bc9152" Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.659001 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:01:15 crc kubenswrapper[4744]: E0106 15:01:15.659676 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/d8242060208b6087b681ac5c3e20e14d67d33eda1997a430d58641069801a50f/diff" to get inode usage: stat /var/lib/containers/storage/overlay/d8242060208b6087b681ac5c3e20e14d67d33eda1997a430d58641069801a50f/diff: no such file or directory, extraDiskErr: Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.770463 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data\") pod \"13f41b35-92ba-4b8a-85cf-969aeccdba04\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.770549 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-combined-ca-bundle\") pod \"13f41b35-92ba-4b8a-85cf-969aeccdba04\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.770748 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8z9d\" (UniqueName: \"kubernetes.io/projected/13f41b35-92ba-4b8a-85cf-969aeccdba04-kube-api-access-r8z9d\") pod \"13f41b35-92ba-4b8a-85cf-969aeccdba04\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.770934 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data-custom\") pod \"13f41b35-92ba-4b8a-85cf-969aeccdba04\" (UID: \"13f41b35-92ba-4b8a-85cf-969aeccdba04\") " Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.780376 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "13f41b35-92ba-4b8a-85cf-969aeccdba04" (UID: "13f41b35-92ba-4b8a-85cf-969aeccdba04"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.788463 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13f41b35-92ba-4b8a-85cf-969aeccdba04-kube-api-access-r8z9d" (OuterVolumeSpecName: "kube-api-access-r8z9d") pod "13f41b35-92ba-4b8a-85cf-969aeccdba04" (UID: "13f41b35-92ba-4b8a-85cf-969aeccdba04"). InnerVolumeSpecName "kube-api-access-r8z9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.807833 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13f41b35-92ba-4b8a-85cf-969aeccdba04" (UID: "13f41b35-92ba-4b8a-85cf-969aeccdba04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:15 crc kubenswrapper[4744]: E0106 15:01:15.828769 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/8f84585f09ac6dad03d2d3ae8b1e9aaa74a35dac7c45838e8f643a7b91ea3cc4/diff" to get inode usage: stat /var/lib/containers/storage/overlay/8f84585f09ac6dad03d2d3ae8b1e9aaa74a35dac7c45838e8f643a7b91ea3cc4/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_neutron-dc6576bbb-qn5wg_826e0141-c6b9-46bc-8450-88cef322d974/neutron-api/0.log" to get inode usage: stat /var/log/pods/openstack_neutron-dc6576bbb-qn5wg_826e0141-c6b9-46bc-8450-88cef322d974/neutron-api/0.log: no such file or directory Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.853068 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data" (OuterVolumeSpecName: "config-data") pod "13f41b35-92ba-4b8a-85cf-969aeccdba04" (UID: "13f41b35-92ba-4b8a-85cf-969aeccdba04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.873893 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.873928 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.873940 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8z9d\" (UniqueName: \"kubernetes.io/projected/13f41b35-92ba-4b8a-85cf-969aeccdba04-kube-api-access-r8z9d\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:15 crc kubenswrapper[4744]: I0106 15:01:15.873948 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/13f41b35-92ba-4b8a-85cf-969aeccdba04-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:16 crc kubenswrapper[4744]: I0106 15:01:16.476877 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerStarted","Data":"2bf2b634b3024dea2befdb1f467ac97467295a00c1502406a72477235fd0e8a4"} Jan 06 15:01:16 crc kubenswrapper[4744]: I0106 15:01:16.478816 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" event={"ID":"13f41b35-92ba-4b8a-85cf-969aeccdba04","Type":"ContainerDied","Data":"b890912e39343e60abdcfe8bf6b589efc3c4b41a586c0bccc7424b18c53a8ef3"} Jan 06 15:01:16 crc kubenswrapper[4744]: I0106 15:01:16.478857 4744 scope.go:117] "RemoveContainer" containerID="a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6" Jan 06 15:01:16 crc kubenswrapper[4744]: I0106 15:01:16.479015 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7bd5fcfc47-lf5fz" Jan 06 15:01:16 crc kubenswrapper[4744]: I0106 15:01:16.567173 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7bd5fcfc47-lf5fz"] Jan 06 15:01:16 crc kubenswrapper[4744]: I0106 15:01:16.582068 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-7bd5fcfc47-lf5fz"] Jan 06 15:01:17 crc kubenswrapper[4744]: E0106 15:01:17.000556 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/46f1f2489e66067a78c5f801b50272539041bde5c33512e2c8a0e249a82bafca/diff" to get inode usage: stat /var/lib/containers/storage/overlay/46f1f2489e66067a78c5f801b50272539041bde5c33512e2c8a0e249a82bafca/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_neutron-dc6576bbb-qn5wg_826e0141-c6b9-46bc-8450-88cef322d974/neutron-httpd/0.log" to get inode usage: stat /var/log/pods/openstack_neutron-dc6576bbb-qn5wg_826e0141-c6b9-46bc-8450-88cef322d974/neutron-httpd/0.log: no such file or directory Jan 06 15:01:17 crc kubenswrapper[4744]: I0106 15:01:17.507686 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8pt" event={"ID":"5405b281-4e69-400c-9702-0ba8b5a20918","Type":"ContainerStarted","Data":"00128f1d60899a2ce9878755347d15779883b1165bac0fcd3f7b2d58d332cb85"} Jan 06 15:01:17 crc kubenswrapper[4744]: I0106 15:01:17.547480 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jf8pt" podStartSLOduration=6.119235549 podStartE2EDuration="28.54746013s" podCreationTimestamp="2026-01-06 15:00:49 +0000 UTC" firstStartedPulling="2026-01-06 15:00:53.972562078 +0000 UTC m=+1450.600028396" lastFinishedPulling="2026-01-06 15:01:16.400786659 +0000 UTC m=+1473.028252977" observedRunningTime="2026-01-06 15:01:17.538800851 +0000 UTC m=+1474.166267169" watchObservedRunningTime="2026-01-06 15:01:17.54746013 +0000 UTC m=+1474.174926448" Jan 06 15:01:17 crc kubenswrapper[4744]: I0106 15:01:17.724262 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13f41b35-92ba-4b8a-85cf-969aeccdba04" path="/var/lib/kubelet/pods/13f41b35-92ba-4b8a-85cf-969aeccdba04/volumes" Jan 06 15:01:18 crc kubenswrapper[4744]: I0106 15:01:18.520045 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerStarted","Data":"d5f05e138cfafdf48e62b274ed67756d3c919ceacba303f614f37f053789fcfb"} Jan 06 15:01:18 crc kubenswrapper[4744]: I0106 15:01:18.520269 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="ceilometer-central-agent" containerID="cri-o://f99ed76641c6b5de4c93710c95b2260d76b028d6b87f58fff59251657ee737fb" gracePeriod=30 Jan 06 15:01:18 crc kubenswrapper[4744]: I0106 15:01:18.520302 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="sg-core" containerID="cri-o://2bf2b634b3024dea2befdb1f467ac97467295a00c1502406a72477235fd0e8a4" gracePeriod=30 Jan 06 15:01:18 crc kubenswrapper[4744]: I0106 15:01:18.520329 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="proxy-httpd" containerID="cri-o://d5f05e138cfafdf48e62b274ed67756d3c919ceacba303f614f37f053789fcfb" gracePeriod=30 Jan 06 15:01:18 crc kubenswrapper[4744]: I0106 15:01:18.520326 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="ceilometer-notification-agent" containerID="cri-o://d365e3d5aee297c5b1288d55446a35840c4d302794f56ad081a607472297daf8" gracePeriod=30 Jan 06 15:01:18 crc kubenswrapper[4744]: I0106 15:01:18.522185 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 15:01:18 crc kubenswrapper[4744]: I0106 15:01:18.558888 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.154627048 podStartE2EDuration="28.558865801s" podCreationTimestamp="2026-01-06 15:00:50 +0000 UTC" firstStartedPulling="2026-01-06 15:00:52.035868805 +0000 UTC m=+1448.663335123" lastFinishedPulling="2026-01-06 15:01:17.440107548 +0000 UTC m=+1474.067573876" observedRunningTime="2026-01-06 15:01:18.546266628 +0000 UTC m=+1475.173732946" watchObservedRunningTime="2026-01-06 15:01:18.558865801 +0000 UTC m=+1475.186332119" Jan 06 15:01:19 crc kubenswrapper[4744]: I0106 15:01:19.543653 4744 generic.go:334] "Generic (PLEG): container finished" podID="d10fd623-6853-4012-8102-649cb3f8fe93" containerID="d5f05e138cfafdf48e62b274ed67756d3c919ceacba303f614f37f053789fcfb" exitCode=0 Jan 06 15:01:19 crc kubenswrapper[4744]: I0106 15:01:19.544232 4744 generic.go:334] "Generic (PLEG): container finished" podID="d10fd623-6853-4012-8102-649cb3f8fe93" containerID="2bf2b634b3024dea2befdb1f467ac97467295a00c1502406a72477235fd0e8a4" exitCode=2 Jan 06 15:01:19 crc kubenswrapper[4744]: I0106 15:01:19.544249 4744 generic.go:334] "Generic (PLEG): container finished" podID="d10fd623-6853-4012-8102-649cb3f8fe93" containerID="f99ed76641c6b5de4c93710c95b2260d76b028d6b87f58fff59251657ee737fb" exitCode=0 Jan 06 15:01:19 crc kubenswrapper[4744]: I0106 15:01:19.543883 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerDied","Data":"d5f05e138cfafdf48e62b274ed67756d3c919ceacba303f614f37f053789fcfb"} Jan 06 15:01:19 crc kubenswrapper[4744]: I0106 15:01:19.544311 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerDied","Data":"2bf2b634b3024dea2befdb1f467ac97467295a00c1502406a72477235fd0e8a4"} Jan 06 15:01:19 crc kubenswrapper[4744]: I0106 15:01:19.544333 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerDied","Data":"f99ed76641c6b5de4c93710c95b2260d76b028d6b87f58fff59251657ee737fb"} Jan 06 15:01:19 crc kubenswrapper[4744]: E0106 15:01:19.997859 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/01063f3c3dc15b8359d141c8d3f2927a762f4014ff344f0808f8dfff3f1cd805/diff" to get inode usage: stat /var/lib/containers/storage/overlay/01063f3c3dc15b8359d141c8d3f2927a762f4014ff344f0808f8dfff3f1cd805/diff: no such file or directory, extraDiskErr: Jan 06 15:01:20 crc kubenswrapper[4744]: I0106 15:01:20.381600 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:01:20 crc kubenswrapper[4744]: I0106 15:01:20.381932 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:01:20 crc kubenswrapper[4744]: E0106 15:01:20.657242 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/3f8a6f7f3c80a955492eb6db467d00560a3daf46f99efa4993b0211215efd8e6/diff" to get inode usage: stat /var/lib/containers/storage/overlay/3f8a6f7f3c80a955492eb6db467d00560a3daf46f99efa4993b0211215efd8e6/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_cinder-api-0_2c546315-1cd0-45e3-809b-c6a8d2e509b3/cinder-api/0.log" to get inode usage: stat /var/log/pods/openstack_cinder-api-0_2c546315-1cd0-45e3-809b-c6a8d2e509b3/cinder-api/0.log: no such file or directory Jan 06 15:01:21 crc kubenswrapper[4744]: E0106 15:01:21.394668 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/65eca0ee3c1c1159b7a506115cbc0be14c25ffe8f9d3f773cf750409cc68b751/diff" to get inode usage: stat /var/lib/containers/storage/overlay/65eca0ee3c1c1159b7a506115cbc0be14c25ffe8f9d3f773cf750409cc68b751/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_c0405b86-e930-436b-94a0-14b08f60f6c1/ceilometer-central-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_c0405b86-e930-436b-94a0-14b08f60f6c1/ceilometer-central-agent/0.log: no such file or directory Jan 06 15:01:21 crc kubenswrapper[4744]: I0106 15:01:21.448977 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jf8pt" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" containerName="registry-server" probeResult="failure" output=< Jan 06 15:01:21 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 15:01:21 crc kubenswrapper[4744]: > Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.589017 4744 generic.go:334] "Generic (PLEG): container finished" podID="d10fd623-6853-4012-8102-649cb3f8fe93" containerID="d365e3d5aee297c5b1288d55446a35840c4d302794f56ad081a607472297daf8" exitCode=0 Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.589581 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerDied","Data":"d365e3d5aee297c5b1288d55446a35840c4d302794f56ad081a607472297daf8"} Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.589608 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d10fd623-6853-4012-8102-649cb3f8fe93","Type":"ContainerDied","Data":"59b9493d3482d8b531877f984e248e6a849ce255a8f382f8304630d4f8424c55"} Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.589619 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59b9493d3482d8b531877f984e248e6a849ce255a8f382f8304630d4f8424c55" Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.684133 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.805074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-sg-core-conf-yaml\") pod \"d10fd623-6853-4012-8102-649cb3f8fe93\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.805142 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-run-httpd\") pod \"d10fd623-6853-4012-8102-649cb3f8fe93\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.805197 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-log-httpd\") pod \"d10fd623-6853-4012-8102-649cb3f8fe93\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.805224 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dbwh\" (UniqueName: \"kubernetes.io/projected/d10fd623-6853-4012-8102-649cb3f8fe93-kube-api-access-4dbwh\") pod \"d10fd623-6853-4012-8102-649cb3f8fe93\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.805334 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-scripts\") pod \"d10fd623-6853-4012-8102-649cb3f8fe93\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.805492 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-config-data\") pod \"d10fd623-6853-4012-8102-649cb3f8fe93\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.805548 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-combined-ca-bundle\") pod \"d10fd623-6853-4012-8102-649cb3f8fe93\" (UID: \"d10fd623-6853-4012-8102-649cb3f8fe93\") " Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.805670 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d10fd623-6853-4012-8102-649cb3f8fe93" (UID: "d10fd623-6853-4012-8102-649cb3f8fe93"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.806190 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.806403 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d10fd623-6853-4012-8102-649cb3f8fe93" (UID: "d10fd623-6853-4012-8102-649cb3f8fe93"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.818432 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-scripts" (OuterVolumeSpecName: "scripts") pod "d10fd623-6853-4012-8102-649cb3f8fe93" (UID: "d10fd623-6853-4012-8102-649cb3f8fe93"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.818471 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d10fd623-6853-4012-8102-649cb3f8fe93-kube-api-access-4dbwh" (OuterVolumeSpecName: "kube-api-access-4dbwh") pod "d10fd623-6853-4012-8102-649cb3f8fe93" (UID: "d10fd623-6853-4012-8102-649cb3f8fe93"). InnerVolumeSpecName "kube-api-access-4dbwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:01:23 crc kubenswrapper[4744]: I0106 15:01:23.844604 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d10fd623-6853-4012-8102-649cb3f8fe93" (UID: "d10fd623-6853-4012-8102-649cb3f8fe93"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:23.908927 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:23.909444 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:23.909466 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dbwh\" (UniqueName: \"kubernetes.io/projected/d10fd623-6853-4012-8102-649cb3f8fe93-kube-api-access-4dbwh\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:23.909477 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d10fd623-6853-4012-8102-649cb3f8fe93-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:23.939214 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-config-data" (OuterVolumeSpecName: "config-data") pod "d10fd623-6853-4012-8102-649cb3f8fe93" (UID: "d10fd623-6853-4012-8102-649cb3f8fe93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:23.958848 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d10fd623-6853-4012-8102-649cb3f8fe93" (UID: "d10fd623-6853-4012-8102-649cb3f8fe93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.011384 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.011421 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d10fd623-6853-4012-8102-649cb3f8fe93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.599306 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.642667 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.654014 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.699183 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:24 crc kubenswrapper[4744]: E0106 15:01:24.700725 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13f41b35-92ba-4b8a-85cf-969aeccdba04" containerName="heat-engine" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.700770 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="13f41b35-92ba-4b8a-85cf-969aeccdba04" containerName="heat-engine" Jan 06 15:01:24 crc kubenswrapper[4744]: E0106 15:01:24.700790 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="ceilometer-central-agent" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.700800 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="ceilometer-central-agent" Jan 06 15:01:24 crc kubenswrapper[4744]: E0106 15:01:24.700822 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="sg-core" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.700832 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="sg-core" Jan 06 15:01:24 crc kubenswrapper[4744]: E0106 15:01:24.700847 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="ceilometer-notification-agent" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.700856 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="ceilometer-notification-agent" Jan 06 15:01:24 crc kubenswrapper[4744]: E0106 15:01:24.700872 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3548869-d949-4d99-8968-8cdf8b347771" containerName="keystone-cron" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.700880 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3548869-d949-4d99-8968-8cdf8b347771" containerName="keystone-cron" Jan 06 15:01:24 crc kubenswrapper[4744]: E0106 15:01:24.700908 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" containerName="heat-api" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.700916 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" containerName="heat-api" Jan 06 15:01:24 crc kubenswrapper[4744]: E0106 15:01:24.700969 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="proxy-httpd" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.700980 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="proxy-httpd" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.701344 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="proxy-httpd" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.701371 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="sg-core" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.701388 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" containerName="heat-api" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.701410 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="ceilometer-notification-agent" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.701421 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="13f41b35-92ba-4b8a-85cf-969aeccdba04" containerName="heat-engine" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.701434 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3548869-d949-4d99-8968-8cdf8b347771" containerName="keystone-cron" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.701449 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" containerName="ceilometer-central-agent" Jan 06 15:01:24 crc kubenswrapper[4744]: E0106 15:01:24.701894 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" containerName="heat-api" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.701922 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" containerName="heat-api" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.702288 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd990c60-8782-4505-8e74-1b99a8b854f2" containerName="heat-api" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.708558 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.717043 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.717382 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.757308 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.832648 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-log-httpd\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.833058 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-scripts\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.833088 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-config-data\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.833201 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-run-httpd\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.833299 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.833342 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2cm9\" (UniqueName: \"kubernetes.io/projected/4bae9d75-869e-406d-be25-81f7e0fcd743-kube-api-access-c2cm9\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.833379 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.935924 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-scripts\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.935989 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-config-data\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.936053 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-run-httpd\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.936122 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.936187 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2cm9\" (UniqueName: \"kubernetes.io/projected/4bae9d75-869e-406d-be25-81f7e0fcd743-kube-api-access-c2cm9\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.936301 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.936372 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-log-httpd\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.936877 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-run-httpd\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.937354 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-log-httpd\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.952147 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-config-data\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.952421 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.952556 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-scripts\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.953888 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:24 crc kubenswrapper[4744]: I0106 15:01:24.967548 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2cm9\" (UniqueName: \"kubernetes.io/projected/4bae9d75-869e-406d-be25-81f7e0fcd743-kube-api-access-c2cm9\") pod \"ceilometer-0\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " pod="openstack/ceilometer-0" Jan 06 15:01:25 crc kubenswrapper[4744]: I0106 15:01:25.048290 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:01:25 crc kubenswrapper[4744]: I0106 15:01:25.675228 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:25 crc kubenswrapper[4744]: I0106 15:01:25.677071 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 15:01:25 crc kubenswrapper[4744]: I0106 15:01:25.730989 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d10fd623-6853-4012-8102-649cb3f8fe93" path="/var/lib/kubelet/pods/d10fd623-6853-4012-8102-649cb3f8fe93/volumes" Jan 06 15:01:26 crc kubenswrapper[4744]: I0106 15:01:26.648048 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerStarted","Data":"764c0ad7f191e3e1ad092e59c123c71fa2313065a62831a3c84ecdbbd0b9dafc"} Jan 06 15:01:26 crc kubenswrapper[4744]: I0106 15:01:26.807749 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:27 crc kubenswrapper[4744]: I0106 15:01:27.664842 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerStarted","Data":"8fd166918b60f542f8a28c6166f4b00521d253c81c0145feb9f0918017d35690"} Jan 06 15:01:28 crc kubenswrapper[4744]: I0106 15:01:28.696634 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" event={"ID":"6eb055cf-7e1b-441b-a1db-4e6479bc9152","Type":"ContainerStarted","Data":"521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664"} Jan 06 15:01:28 crc kubenswrapper[4744]: I0106 15:01:28.700136 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerStarted","Data":"ab2c367aad1628660b8b2b22e9de7cc4e7252b4e81c6e96ffee911f3cb3636ea"} Jan 06 15:01:28 crc kubenswrapper[4744]: I0106 15:01:28.739582 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" podStartSLOduration=2.203730103 podStartE2EDuration="42.739541582s" podCreationTimestamp="2026-01-06 15:00:46 +0000 UTC" firstStartedPulling="2026-01-06 15:00:47.665787864 +0000 UTC m=+1444.293254172" lastFinishedPulling="2026-01-06 15:01:28.201599333 +0000 UTC m=+1484.829065651" observedRunningTime="2026-01-06 15:01:28.727918625 +0000 UTC m=+1485.355384973" watchObservedRunningTime="2026-01-06 15:01:28.739541582 +0000 UTC m=+1485.367007890" Jan 06 15:01:30 crc kubenswrapper[4744]: I0106 15:01:30.470282 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:01:30 crc kubenswrapper[4744]: I0106 15:01:30.553794 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:01:30 crc kubenswrapper[4744]: I0106 15:01:30.726645 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jf8pt"] Jan 06 15:01:31 crc kubenswrapper[4744]: I0106 15:01:31.731126 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jf8pt" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" containerName="registry-server" containerID="cri-o://00128f1d60899a2ce9878755347d15779883b1165bac0fcd3f7b2d58d332cb85" gracePeriod=2 Jan 06 15:01:34 crc kubenswrapper[4744]: E0106 15:01:34.175392 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/29a96b2112da9692f83195c1189d525ad07276015ed8be8e3d08618c9ea0ac1c/diff" to get inode usage: stat /var/lib/containers/storage/overlay/29a96b2112da9692f83195c1189d525ad07276015ed8be8e3d08618c9ea0ac1c/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_c0405b86-e930-436b-94a0-14b08f60f6c1/ceilometer-notification-agent/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_c0405b86-e930-436b-94a0-14b08f60f6c1/ceilometer-notification-agent/0.log: no such file or directory Jan 06 15:01:34 crc kubenswrapper[4744]: E0106 15:01:34.333222 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/970ea61945031dc0a9b7fe0271296af00feecccde0aa3b7b27a12816bd4656d4/diff" to get inode usage: stat /var/lib/containers/storage/overlay/970ea61945031dc0a9b7fe0271296af00feecccde0aa3b7b27a12816bd4656d4/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_heat-engine-7bd5fcfc47-lf5fz_13f41b35-92ba-4b8a-85cf-969aeccdba04/heat-engine/0.log" to get inode usage: stat /var/log/pods/openstack_heat-engine-7bd5fcfc47-lf5fz_13f41b35-92ba-4b8a-85cf-969aeccdba04/heat-engine/0.log: no such file or directory Jan 06 15:01:34 crc kubenswrapper[4744]: I0106 15:01:34.489182 4744 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 15:01:34 crc kubenswrapper[4744]: I0106 15:01:34.489304 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.58:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 06 15:01:35 crc kubenswrapper[4744]: E0106 15:01:35.108287 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/e6b4e74b161c5b23c9a28da647ef5882a9e0c2f577ba9696b131e6191a011975/diff" to get inode usage: stat /var/lib/containers/storage/overlay/e6b4e74b161c5b23c9a28da647ef5882a9e0c2f577ba9696b131e6191a011975/diff: no such file or directory, extraDiskErr: Jan 06 15:01:35 crc kubenswrapper[4744]: E0106 15:01:35.536746 4744 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/d0600a50f79d05396af8af0eaf13fa5c1332d9f677531953014cf08f10930fe3/diff" to get inode usage: stat /var/lib/containers/storage/overlay/d0600a50f79d05396af8af0eaf13fa5c1332d9f677531953014cf08f10930fe3/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_ceilometer-0_c0405b86-e930-436b-94a0-14b08f60f6c1/sg-core/0.log" to get inode usage: stat /var/log/pods/openstack_ceilometer-0_c0405b86-e930-436b-94a0-14b08f60f6c1/sg-core/0.log: no such file or directory Jan 06 15:01:37 crc kubenswrapper[4744]: I0106 15:01:37.847717 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jf8pt_5405b281-4e69-400c-9702-0ba8b5a20918/registry-server/0.log" Jan 06 15:01:37 crc kubenswrapper[4744]: I0106 15:01:37.855433 4744 generic.go:334] "Generic (PLEG): container finished" podID="5405b281-4e69-400c-9702-0ba8b5a20918" containerID="00128f1d60899a2ce9878755347d15779883b1165bac0fcd3f7b2d58d332cb85" exitCode=137 Jan 06 15:01:37 crc kubenswrapper[4744]: I0106 15:01:37.855575 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8pt" event={"ID":"5405b281-4e69-400c-9702-0ba8b5a20918","Type":"ContainerDied","Data":"00128f1d60899a2ce9878755347d15779883b1165bac0fcd3f7b2d58d332cb85"} Jan 06 15:01:37 crc kubenswrapper[4744]: I0106 15:01:37.912283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerStarted","Data":"d07a66535886dc00388823963c0bc3ee70e6ff5c0c54ed7d63f4947b094685ab"} Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.110240 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jf8pt_5405b281-4e69-400c-9702-0ba8b5a20918/registry-server/0.log" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.111847 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.268170 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-catalog-content\") pod \"5405b281-4e69-400c-9702-0ba8b5a20918\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.268558 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-utilities\") pod \"5405b281-4e69-400c-9702-0ba8b5a20918\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.268583 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt6k6\" (UniqueName: \"kubernetes.io/projected/5405b281-4e69-400c-9702-0ba8b5a20918-kube-api-access-gt6k6\") pod \"5405b281-4e69-400c-9702-0ba8b5a20918\" (UID: \"5405b281-4e69-400c-9702-0ba8b5a20918\") " Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.271834 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-utilities" (OuterVolumeSpecName: "utilities") pod "5405b281-4e69-400c-9702-0ba8b5a20918" (UID: "5405b281-4e69-400c-9702-0ba8b5a20918"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.282437 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5405b281-4e69-400c-9702-0ba8b5a20918-kube-api-access-gt6k6" (OuterVolumeSpecName: "kube-api-access-gt6k6") pod "5405b281-4e69-400c-9702-0ba8b5a20918" (UID: "5405b281-4e69-400c-9702-0ba8b5a20918"). InnerVolumeSpecName "kube-api-access-gt6k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.324111 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5405b281-4e69-400c-9702-0ba8b5a20918" (UID: "5405b281-4e69-400c-9702-0ba8b5a20918"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.371485 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.371533 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5405b281-4e69-400c-9702-0ba8b5a20918-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.371547 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt6k6\" (UniqueName: \"kubernetes.io/projected/5405b281-4e69-400c-9702-0ba8b5a20918-kube-api-access-gt6k6\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.924728 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-jf8pt_5405b281-4e69-400c-9702-0ba8b5a20918/registry-server/0.log" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.925774 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jf8pt" event={"ID":"5405b281-4e69-400c-9702-0ba8b5a20918","Type":"ContainerDied","Data":"b8b01bdb832da24a924dd91c5bbf268f31e96becaaa4821ea03fd87939105af1"} Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.925817 4744 scope.go:117] "RemoveContainer" containerID="00128f1d60899a2ce9878755347d15779883b1165bac0fcd3f7b2d58d332cb85" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.925865 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jf8pt" Jan 06 15:01:38 crc kubenswrapper[4744]: I0106 15:01:38.989890 4744 scope.go:117] "RemoveContainer" containerID="7c16cea3773de90e3b6682b5c0afa4a55dfbc394fa88316d7d785a2557cb6a60" Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.002422 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jf8pt"] Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.023517 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jf8pt"] Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.046382 4744 scope.go:117] "RemoveContainer" containerID="84990ca01c1f1c6bf56e04da9e76ed95e4c33723a462ee9c5cc77e6dec6a2d0b" Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.728883 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" path="/var/lib/kubelet/pods/5405b281-4e69-400c-9702-0ba8b5a20918/volumes" Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.941058 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerStarted","Data":"33b3c55c42a07f503c76242eeef8a8523ba29f0d9624d14b8c649b8c088bad23"} Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.941310 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="ceilometer-central-agent" containerID="cri-o://8fd166918b60f542f8a28c6166f4b00521d253c81c0145feb9f0918017d35690" gracePeriod=30 Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.941571 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.941962 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="proxy-httpd" containerID="cri-o://33b3c55c42a07f503c76242eeef8a8523ba29f0d9624d14b8c649b8c088bad23" gracePeriod=30 Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.942016 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="sg-core" containerID="cri-o://d07a66535886dc00388823963c0bc3ee70e6ff5c0c54ed7d63f4947b094685ab" gracePeriod=30 Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.942058 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="ceilometer-notification-agent" containerID="cri-o://ab2c367aad1628660b8b2b22e9de7cc4e7252b4e81c6e96ffee911f3cb3636ea" gracePeriod=30 Jan 06 15:01:39 crc kubenswrapper[4744]: I0106 15:01:39.986385 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.249642048 podStartE2EDuration="15.986366754s" podCreationTimestamp="2026-01-06 15:01:24 +0000 UTC" firstStartedPulling="2026-01-06 15:01:25.676687822 +0000 UTC m=+1482.304154140" lastFinishedPulling="2026-01-06 15:01:39.413412518 +0000 UTC m=+1496.040878846" observedRunningTime="2026-01-06 15:01:39.984497204 +0000 UTC m=+1496.611963532" watchObservedRunningTime="2026-01-06 15:01:39.986366754 +0000 UTC m=+1496.613833072" Jan 06 15:01:40 crc kubenswrapper[4744]: I0106 15:01:40.955016 4744 generic.go:334] "Generic (PLEG): container finished" podID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerID="33b3c55c42a07f503c76242eeef8a8523ba29f0d9624d14b8c649b8c088bad23" exitCode=0 Jan 06 15:01:40 crc kubenswrapper[4744]: I0106 15:01:40.955370 4744 generic.go:334] "Generic (PLEG): container finished" podID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerID="d07a66535886dc00388823963c0bc3ee70e6ff5c0c54ed7d63f4947b094685ab" exitCode=2 Jan 06 15:01:40 crc kubenswrapper[4744]: I0106 15:01:40.955382 4744 generic.go:334] "Generic (PLEG): container finished" podID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerID="ab2c367aad1628660b8b2b22e9de7cc4e7252b4e81c6e96ffee911f3cb3636ea" exitCode=0 Jan 06 15:01:40 crc kubenswrapper[4744]: I0106 15:01:40.955097 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerDied","Data":"33b3c55c42a07f503c76242eeef8a8523ba29f0d9624d14b8c649b8c088bad23"} Jan 06 15:01:40 crc kubenswrapper[4744]: I0106 15:01:40.955449 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerDied","Data":"d07a66535886dc00388823963c0bc3ee70e6ff5c0c54ed7d63f4947b094685ab"} Jan 06 15:01:40 crc kubenswrapper[4744]: I0106 15:01:40.955471 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerDied","Data":"ab2c367aad1628660b8b2b22e9de7cc4e7252b4e81c6e96ffee911f3cb3636ea"} Jan 06 15:01:44 crc kubenswrapper[4744]: I0106 15:01:44.423940 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:01:44 crc kubenswrapper[4744]: I0106 15:01:44.424881 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:01:44 crc kubenswrapper[4744]: I0106 15:01:44.424954 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:01:44 crc kubenswrapper[4744]: I0106 15:01:44.426591 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:01:44 crc kubenswrapper[4744]: I0106 15:01:44.426664 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" gracePeriod=600 Jan 06 15:01:44 crc kubenswrapper[4744]: E0106 15:01:44.573918 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:01:45 crc kubenswrapper[4744]: I0106 15:01:45.008700 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" exitCode=0 Jan 06 15:01:45 crc kubenswrapper[4744]: I0106 15:01:45.008738 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c"} Jan 06 15:01:45 crc kubenswrapper[4744]: I0106 15:01:45.008794 4744 scope.go:117] "RemoveContainer" containerID="8cc1261767d494e24246d9d799ed9782619069fb6a47b29f76fd6dc783322bbb" Jan 06 15:01:45 crc kubenswrapper[4744]: I0106 15:01:45.009562 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:01:45 crc kubenswrapper[4744]: E0106 15:01:45.009826 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.680205 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2.scope WatchSource:0}: Error finding container c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2: Status 404 returned error can't find the container with id c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2 Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.703707 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice/crio-49837ba44b56a275f99b5cb68926b5bfb9547afbe1b81a6ee5d41e9b64b6362e.scope WatchSource:0}: Error finding container 49837ba44b56a275f99b5cb68926b5bfb9547afbe1b81a6ee5d41e9b64b6362e: Status 404 returned error can't find the container with id 49837ba44b56a275f99b5cb68926b5bfb9547afbe1b81a6ee5d41e9b64b6362e Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.704383 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd990c60_8782_4505_8e74_1b99a8b854f2.slice/crio-305ac2710be228bb6cbc8afbc58b3363f2919bbbb1081a2f7409ccb9e4412142 WatchSource:0}: Error finding container 305ac2710be228bb6cbc8afbc58b3363f2919bbbb1081a2f7409ccb9e4412142: Status 404 returned error can't find the container with id 305ac2710be228bb6cbc8afbc58b3363f2919bbbb1081a2f7409ccb9e4412142 Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.710195 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd990c60_8782_4505_8e74_1b99a8b854f2.slice/crio-92dcbb933ea1350c3a509c4e87a05745b2c1e7c2b0345127e1ace355bdd2ef51.scope WatchSource:0}: Error finding container 92dcbb933ea1350c3a509c4e87a05745b2c1e7c2b0345127e1ace355bdd2ef51: Status 404 returned error can't find the container with id 92dcbb933ea1350c3a509c4e87a05745b2c1e7c2b0345127e1ace355bdd2ef51 Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.715057 4744 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5405b281_4e69_400c_9702_0ba8b5a20918.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5405b281_4e69_400c_9702_0ba8b5a20918.slice: no such file or directory Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.721349 4744 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd10fd623_6853_4012_8102_649cb3f8fe93.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd10fd623_6853_4012_8102_649cb3f8fe93.slice: no such file or directory Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.731199 4744 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice/crio-conmon-4934cb90dabe164007c541855cc13b4c1e084cb7d66a8f9ed70cee5a95f93e41.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice/crio-conmon-4934cb90dabe164007c541855cc13b4c1e084cb7d66a8f9ed70cee5a95f93e41.scope: no such file or directory Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.731337 4744 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd990c60_8782_4505_8e74_1b99a8b854f2.slice/crio-conmon-03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd990c60_8782_4505_8e74_1b99a8b854f2.slice/crio-conmon-03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603.scope: no such file or directory Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.731378 4744 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice/crio-4934cb90dabe164007c541855cc13b4c1e084cb7d66a8f9ed70cee5a95f93e41.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice/crio-4934cb90dabe164007c541855cc13b4c1e084cb7d66a8f9ed70cee5a95f93e41.scope: no such file or directory Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.731399 4744 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd990c60_8782_4505_8e74_1b99a8b854f2.slice/crio-03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd990c60_8782_4505_8e74_1b99a8b854f2.slice/crio-03af2267e6d31411823cb8a6c78a9692e32a7b3e996c99df956904e1fa81c603.scope: no such file or directory Jan 06 15:01:45 crc kubenswrapper[4744]: W0106 15:01:45.732173 4744 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3548869_d949_4d99_8968_8cdf8b347771.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3548869_d949_4d99_8968_8cdf8b347771.slice: no such file or directory Jan 06 15:01:45 crc kubenswrapper[4744]: E0106 15:01:45.807878 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice/crio-ac3f5244aed3dee21656335f51e76d8dae8302b455cd4d0d22134f9d9bc7efd6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf6d608b_355b_471e_9fab_40b62a25b8f0.slice/crio-6fb418c65bf1ed1af72f36bd8c4cb20d78214bcc63a2c286c95b3c5da6c36dc5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice/crio-a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-conmon-9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-7a1eabb1fd46fa9eaa70c0863168b519b6bb07624f9e45e231fc57023afbe68d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-conmon-6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd990c60_8782_4505_8e74_1b99a8b854f2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf6d608b_355b_471e_9fab_40b62a25b8f0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-2b06a91900b437fe3ccafd75d3bf8483ade95942fcff3763072c6fbcdd086cf4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-82433d8804f062db0a8309b9a32c6eeaa91a4c5e5b9743629079375e902c1cae\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice/crio-b890912e39343e60abdcfe8bf6b589efc3c4b41a586c0bccc7424b18c53a8ef3\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice/crio-conmon-a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-conmon-311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice\": RecentStats: unable to find data in memory cache]" Jan 06 15:01:45 crc kubenswrapper[4744]: E0106 15:01:45.808071 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice/crio-conmon-a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-7a1eabb1fd46fa9eaa70c0863168b519b6bb07624f9e45e231fc57023afbe68d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-2b06a91900b437fe3ccafd75d3bf8483ade95942fcff3763072c6fbcdd086cf4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-conmon-311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice/crio-a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice/crio-b890912e39343e60abdcfe8bf6b589efc3c4b41a586c0bccc7424b18c53a8ef3\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-82433d8804f062db0a8309b9a32c6eeaa91a4c5e5b9743629079375e902c1cae\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-002d4b36635e093b1fa011fd8aa57466323ff773fe56e2f8fd80bfc94a5305aa\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-conmon-9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf6d608b_355b_471e_9fab_40b62a25b8f0.slice/crio-6fb418c65bf1ed1af72f36bd8c4cb20d78214bcc63a2c286c95b3c5da6c36dc5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf6d608b_355b_471e_9fab_40b62a25b8f0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-conmon-6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-conmon-4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice/crio-ac3f5244aed3dee21656335f51e76d8dae8302b455cd4d0d22134f9d9bc7efd6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd990c60_8782_4505_8e74_1b99a8b854f2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice\": RecentStats: unable to find data in memory cache]" Jan 06 15:01:45 crc kubenswrapper[4744]: E0106 15:01:45.809988 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf6d608b_355b_471e_9fab_40b62a25b8f0.slice/crio-6fb418c65bf1ed1af72f36bd8c4cb20d78214bcc63a2c286c95b3c5da6c36dc5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-82433d8804f062db0a8309b9a32c6eeaa91a4c5e5b9743629079375e902c1cae\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-conmon-9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-conmon-4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-7a1eabb1fd46fa9eaa70c0863168b519b6bb07624f9e45e231fc57023afbe68d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-bf1a09ffc5d0431897beada6e8b178654802d46049cf681f2da389e3f5fae36f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice/crio-a40237a364b454f7cc5e7d12e7848b82841be743e569ab39caafc503aeae4ce6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-a9304c4da3870dad85ac5b819979489f667cacb2ee4465db04d0d59c2ffe23a6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice/crio-b890912e39343e60abdcfe8bf6b589efc3c4b41a586c0bccc7424b18c53a8ef3\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-c29e10ca647f2db8d8a9006451419591d5d2e480583f134ae39d30c3c43604a2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd990c60_8782_4505_8e74_1b99a8b854f2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf6d608b_355b_471e_9fab_40b62a25b8f0.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-conmon-67b96380b2afb08f3b5975f5a4ddcd3ca406300bf910956a2be612b1e3a98aa9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice/crio-9a1e01ee13284399cc2c307df3e76bfa68ae02e4d766ca0d9567df03c001efea.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c546315_1cd0_45e3_809b_c6a8d2e509b3.slice/crio-conmon-6d30704b23f972753547716fe2e6f4049fbc10a66406a8a0a198fff45c33c0f9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice/crio-ac3f5244aed3dee21656335f51e76d8dae8302b455cd4d0d22134f9d9bc7efd6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-4bdbdcfb664cc60364f1b78943e626f07a5a68f074a08c511a0b792ce93377f3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-conmon-311d8f1ea2052f6e25544440360941c50d068de4adf22a601fa52c810b957489.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13f41b35_92ba_4b8a_85cf_969aeccdba04.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod826e0141_c6b9_46bc_8450_88cef322d974.slice/crio-2b06a91900b437fe3ccafd75d3bf8483ade95942fcff3763072c6fbcdd086cf4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e405ef_b09a_47b9_902e_db0ff2df7788.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0405b86_e930_436b_94a0_14b08f60f6c1.slice/crio-002d4b36635e093b1fa011fd8aa57466323ff773fe56e2f8fd80bfc94a5305aa\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcce76110_c2be_496e_92ff_b585bde98eb8.slice\": RecentStats: unable to find data in memory cache]" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.023970 4744 generic.go:334] "Generic (PLEG): container finished" podID="aefa115f-c019-4d47-be4c-659e19fe9eea" containerID="3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a" exitCode=137 Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.024074 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69c8df499f-wjrkv" event={"ID":"aefa115f-c019-4d47-be4c-659e19fe9eea","Type":"ContainerDied","Data":"3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a"} Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.024181 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69c8df499f-wjrkv" event={"ID":"aefa115f-c019-4d47-be4c-659e19fe9eea","Type":"ContainerDied","Data":"ed2df8acb6444d1b71421733b7082c76977cdb3527c193e5cd9910c5c9dbb63c"} Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.024221 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed2df8acb6444d1b71421733b7082c76977cdb3527c193e5cd9910c5c9dbb63c" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.115422 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.122229 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm526\" (UniqueName: \"kubernetes.io/projected/aefa115f-c019-4d47-be4c-659e19fe9eea-kube-api-access-dm526\") pod \"aefa115f-c019-4d47-be4c-659e19fe9eea\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.122439 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data-custom\") pod \"aefa115f-c019-4d47-be4c-659e19fe9eea\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.128083 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aefa115f-c019-4d47-be4c-659e19fe9eea-kube-api-access-dm526" (OuterVolumeSpecName: "kube-api-access-dm526") pod "aefa115f-c019-4d47-be4c-659e19fe9eea" (UID: "aefa115f-c019-4d47-be4c-659e19fe9eea"). InnerVolumeSpecName "kube-api-access-dm526". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.133576 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "aefa115f-c019-4d47-be4c-659e19fe9eea" (UID: "aefa115f-c019-4d47-be4c-659e19fe9eea"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.225896 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-combined-ca-bundle\") pod \"aefa115f-c019-4d47-be4c-659e19fe9eea\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.226060 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data\") pod \"aefa115f-c019-4d47-be4c-659e19fe9eea\" (UID: \"aefa115f-c019-4d47-be4c-659e19fe9eea\") " Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.226667 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.226683 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm526\" (UniqueName: \"kubernetes.io/projected/aefa115f-c019-4d47-be4c-659e19fe9eea-kube-api-access-dm526\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.262252 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aefa115f-c019-4d47-be4c-659e19fe9eea" (UID: "aefa115f-c019-4d47-be4c-659e19fe9eea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.334873 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.335942 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data" (OuterVolumeSpecName: "config-data") pod "aefa115f-c019-4d47-be4c-659e19fe9eea" (UID: "aefa115f-c019-4d47-be4c-659e19fe9eea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:46 crc kubenswrapper[4744]: I0106 15:01:46.437221 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aefa115f-c019-4d47-be4c-659e19fe9eea-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:47 crc kubenswrapper[4744]: I0106 15:01:47.039097 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69c8df499f-wjrkv" Jan 06 15:01:47 crc kubenswrapper[4744]: I0106 15:01:47.075793 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-69c8df499f-wjrkv"] Jan 06 15:01:47 crc kubenswrapper[4744]: I0106 15:01:47.087888 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-69c8df499f-wjrkv"] Jan 06 15:01:47 crc kubenswrapper[4744]: I0106 15:01:47.727456 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aefa115f-c019-4d47-be4c-659e19fe9eea" path="/var/lib/kubelet/pods/aefa115f-c019-4d47-be4c-659e19fe9eea/volumes" Jan 06 15:01:50 crc kubenswrapper[4744]: E0106 15:01:50.800062 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb055cf_7e1b_441b_a1db_4e6479bc9152.slice/crio-conmon-521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb055cf_7e1b_441b_a1db_4e6479bc9152.slice/crio-521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-conmon-3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-ed2df8acb6444d1b71421733b7082c76977cdb3527c193e5cd9910c5c9dbb63c\": RecentStats: unable to find data in memory cache]" Jan 06 15:01:50 crc kubenswrapper[4744]: E0106 15:01:50.800953 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-ed2df8acb6444d1b71421733b7082c76977cdb3527c193e5cd9910c5c9dbb63c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1639b17c_035d_4955_a3b6_ac33889e1c89.slice/crio-conmon-2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb055cf_7e1b_441b_a1db_4e6479bc9152.slice/crio-521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-conmon-3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1639b17c_035d_4955_a3b6_ac33889e1c89.slice/crio-2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bae9d75_869e_406d_be25_81f7e0fcd743.slice/crio-8fd166918b60f542f8a28c6166f4b00521d253c81c0145feb9f0918017d35690.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb055cf_7e1b_441b_a1db_4e6479bc9152.slice/crio-conmon-521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:01:50 crc kubenswrapper[4744]: E0106 15:01:50.808446 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1639b17c_035d_4955_a3b6_ac33889e1c89.slice/crio-conmon-2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb055cf_7e1b_441b_a1db_4e6479bc9152.slice/crio-conmon-521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-conmon-3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bae9d75_869e_406d_be25_81f7e0fcd743.slice/crio-8fd166918b60f542f8a28c6166f4b00521d253c81c0145feb9f0918017d35690.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6eb055cf_7e1b_441b_a1db_4e6479bc9152.slice/crio-521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaefa115f_c019_4d47_be4c_659e19fe9eea.slice/crio-ed2df8acb6444d1b71421733b7082c76977cdb3527c193e5cd9910c5c9dbb63c\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1639b17c_035d_4955_a3b6_ac33889e1c89.slice/crio-2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:01:51 crc kubenswrapper[4744]: I0106 15:01:51.079439 4744 generic.go:334] "Generic (PLEG): container finished" podID="6eb055cf-7e1b-441b-a1db-4e6479bc9152" containerID="521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664" exitCode=0 Jan 06 15:01:51 crc kubenswrapper[4744]: I0106 15:01:51.079514 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" event={"ID":"6eb055cf-7e1b-441b-a1db-4e6479bc9152","Type":"ContainerDied","Data":"521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664"} Jan 06 15:01:51 crc kubenswrapper[4744]: I0106 15:01:51.083361 4744 generic.go:334] "Generic (PLEG): container finished" podID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerID="8fd166918b60f542f8a28c6166f4b00521d253c81c0145feb9f0918017d35690" exitCode=0 Jan 06 15:01:51 crc kubenswrapper[4744]: I0106 15:01:51.083410 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerDied","Data":"8fd166918b60f542f8a28c6166f4b00521d253c81c0145feb9f0918017d35690"} Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.692257 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.873489 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-log-httpd\") pod \"4bae9d75-869e-406d-be25-81f7e0fcd743\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.873554 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-config-data\") pod \"4bae9d75-869e-406d-be25-81f7e0fcd743\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.873837 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-sg-core-conf-yaml\") pod \"4bae9d75-869e-406d-be25-81f7e0fcd743\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.873929 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2cm9\" (UniqueName: \"kubernetes.io/projected/4bae9d75-869e-406d-be25-81f7e0fcd743-kube-api-access-c2cm9\") pod \"4bae9d75-869e-406d-be25-81f7e0fcd743\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.874002 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-scripts\") pod \"4bae9d75-869e-406d-be25-81f7e0fcd743\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.874066 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-combined-ca-bundle\") pod \"4bae9d75-869e-406d-be25-81f7e0fcd743\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.874093 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-run-httpd\") pod \"4bae9d75-869e-406d-be25-81f7e0fcd743\" (UID: \"4bae9d75-869e-406d-be25-81f7e0fcd743\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.874558 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4bae9d75-869e-406d-be25-81f7e0fcd743" (UID: "4bae9d75-869e-406d-be25-81f7e0fcd743"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.874595 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4bae9d75-869e-406d-be25-81f7e0fcd743" (UID: "4bae9d75-869e-406d-be25-81f7e0fcd743"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.875131 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.875148 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4bae9d75-869e-406d-be25-81f7e0fcd743-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.882357 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-scripts" (OuterVolumeSpecName: "scripts") pod "4bae9d75-869e-406d-be25-81f7e0fcd743" (UID: "4bae9d75-869e-406d-be25-81f7e0fcd743"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.888030 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bae9d75-869e-406d-be25-81f7e0fcd743-kube-api-access-c2cm9" (OuterVolumeSpecName: "kube-api-access-c2cm9") pod "4bae9d75-869e-406d-be25-81f7e0fcd743" (UID: "4bae9d75-869e-406d-be25-81f7e0fcd743"). InnerVolumeSpecName "kube-api-access-c2cm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.906101 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4bae9d75-869e-406d-be25-81f7e0fcd743" (UID: "4bae9d75-869e-406d-be25-81f7e0fcd743"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.977782 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.977815 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2cm9\" (UniqueName: \"kubernetes.io/projected/4bae9d75-869e-406d-be25-81f7e0fcd743-kube-api-access-c2cm9\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:51.977887 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.000721 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bae9d75-869e-406d-be25-81f7e0fcd743" (UID: "4bae9d75-869e-406d-be25-81f7e0fcd743"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.030539 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-config-data" (OuterVolumeSpecName: "config-data") pod "4bae9d75-869e-406d-be25-81f7e0fcd743" (UID: "4bae9d75-869e-406d-be25-81f7e0fcd743"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.079011 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.079054 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bae9d75-869e-406d-be25-81f7e0fcd743-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.101263 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.103219 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4bae9d75-869e-406d-be25-81f7e0fcd743","Type":"ContainerDied","Data":"764c0ad7f191e3e1ad092e59c123c71fa2313065a62831a3c84ecdbbd0b9dafc"} Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.103294 4744 scope.go:117] "RemoveContainer" containerID="33b3c55c42a07f503c76242eeef8a8523ba29f0d9624d14b8c649b8c088bad23" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.176202 4744 scope.go:117] "RemoveContainer" containerID="d07a66535886dc00388823963c0bc3ee70e6ff5c0c54ed7d63f4947b094685ab" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.184531 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.209141 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.215361 4744 scope.go:117] "RemoveContainer" containerID="ab2c367aad1628660b8b2b22e9de7cc4e7252b4e81c6e96ffee911f3cb3636ea" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.227704 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:53 crc kubenswrapper[4744]: E0106 15:01:52.228420 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefa115f-c019-4d47-be4c-659e19fe9eea" containerName="heat-api" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228435 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefa115f-c019-4d47-be4c-659e19fe9eea" containerName="heat-api" Jan 06 15:01:53 crc kubenswrapper[4744]: E0106 15:01:52.228457 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="sg-core" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228463 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="sg-core" Jan 06 15:01:53 crc kubenswrapper[4744]: E0106 15:01:52.228493 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" containerName="registry-server" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228499 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" containerName="registry-server" Jan 06 15:01:53 crc kubenswrapper[4744]: E0106 15:01:52.228513 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" containerName="extract-utilities" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228519 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" containerName="extract-utilities" Jan 06 15:01:53 crc kubenswrapper[4744]: E0106 15:01:52.228527 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" containerName="extract-content" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228533 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" containerName="extract-content" Jan 06 15:01:53 crc kubenswrapper[4744]: E0106 15:01:52.228545 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="proxy-httpd" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228550 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="proxy-httpd" Jan 06 15:01:53 crc kubenswrapper[4744]: E0106 15:01:52.228564 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="ceilometer-notification-agent" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228569 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="ceilometer-notification-agent" Jan 06 15:01:53 crc kubenswrapper[4744]: E0106 15:01:52.228579 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="ceilometer-central-agent" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228585 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="ceilometer-central-agent" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228774 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="proxy-httpd" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228786 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="sg-core" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228796 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5405b281-4e69-400c-9702-0ba8b5a20918" containerName="registry-server" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228803 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="ceilometer-notification-agent" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228819 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="aefa115f-c019-4d47-be4c-659e19fe9eea" containerName="heat-api" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.228827 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" containerName="ceilometer-central-agent" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.233979 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.236280 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.238634 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.254348 4744 scope.go:117] "RemoveContainer" containerID="8fd166918b60f542f8a28c6166f4b00521d253c81c0145feb9f0918017d35690" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.336829 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.384558 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-run-httpd\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.384653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.384700 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggz5t\" (UniqueName: \"kubernetes.io/projected/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-kube-api-access-ggz5t\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.384800 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.385019 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-config-data\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.385085 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-log-httpd\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.385111 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-scripts\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.487271 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggz5t\" (UniqueName: \"kubernetes.io/projected/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-kube-api-access-ggz5t\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.487362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.487417 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-config-data\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.487450 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-log-httpd\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.487469 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-scripts\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.487527 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-run-httpd\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.487566 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.488073 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-log-httpd\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.488089 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-run-httpd\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.493575 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-scripts\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.496648 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-config-data\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.496907 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.500977 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.520955 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggz5t\" (UniqueName: \"kubernetes.io/projected/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-kube-api-access-ggz5t\") pod \"ceilometer-0\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:52.550185 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.594281 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.620362 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-config-data\") pod \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.620436 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-scripts\") pod \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.620476 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7b68\" (UniqueName: \"kubernetes.io/projected/6eb055cf-7e1b-441b-a1db-4e6479bc9152-kube-api-access-c7b68\") pod \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.620689 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-combined-ca-bundle\") pod \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\" (UID: \"6eb055cf-7e1b-441b-a1db-4e6479bc9152\") " Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.628738 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eb055cf-7e1b-441b-a1db-4e6479bc9152-kube-api-access-c7b68" (OuterVolumeSpecName: "kube-api-access-c7b68") pod "6eb055cf-7e1b-441b-a1db-4e6479bc9152" (UID: "6eb055cf-7e1b-441b-a1db-4e6479bc9152"). InnerVolumeSpecName "kube-api-access-c7b68". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.630581 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-scripts" (OuterVolumeSpecName: "scripts") pod "6eb055cf-7e1b-441b-a1db-4e6479bc9152" (UID: "6eb055cf-7e1b-441b-a1db-4e6479bc9152"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.672736 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-config-data" (OuterVolumeSpecName: "config-data") pod "6eb055cf-7e1b-441b-a1db-4e6479bc9152" (UID: "6eb055cf-7e1b-441b-a1db-4e6479bc9152"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.701324 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6eb055cf-7e1b-441b-a1db-4e6479bc9152" (UID: "6eb055cf-7e1b-441b-a1db-4e6479bc9152"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.725074 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.725141 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.725182 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7b68\" (UniqueName: \"kubernetes.io/projected/6eb055cf-7e1b-441b-a1db-4e6479bc9152-kube-api-access-c7b68\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.725198 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6eb055cf-7e1b-441b-a1db-4e6479bc9152-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:01:53 crc kubenswrapper[4744]: W0106 15:01:53.733303 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0317db72_7dd0_4a5f_b1db_b8d79cc9d39a.slice/crio-2165586d9d67fa9714a167052f13c402648746a4b4793f4ae01b6c3344650f17 WatchSource:0}: Error finding container 2165586d9d67fa9714a167052f13c402648746a4b4793f4ae01b6c3344650f17: Status 404 returned error can't find the container with id 2165586d9d67fa9714a167052f13c402648746a4b4793f4ae01b6c3344650f17 Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.752426 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bae9d75-869e-406d-be25-81f7e0fcd743" path="/var/lib/kubelet/pods/4bae9d75-869e-406d-be25-81f7e0fcd743/volumes" Jan 06 15:01:53 crc kubenswrapper[4744]: I0106 15:01:53.755967 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.143502 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" event={"ID":"6eb055cf-7e1b-441b-a1db-4e6479bc9152","Type":"ContainerDied","Data":"07d7405a66bac66f04c991feb98a52f6f30bce6fcaf61e1fd1689a56caedf024"} Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.143555 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07d7405a66bac66f04c991feb98a52f6f30bce6fcaf61e1fd1689a56caedf024" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.143662 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gb9d6" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.146413 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerStarted","Data":"2165586d9d67fa9714a167052f13c402648746a4b4793f4ae01b6c3344650f17"} Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.740219 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 06 15:01:54 crc kubenswrapper[4744]: E0106 15:01:54.741451 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eb055cf-7e1b-441b-a1db-4e6479bc9152" containerName="nova-cell0-conductor-db-sync" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.741472 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eb055cf-7e1b-441b-a1db-4e6479bc9152" containerName="nova-cell0-conductor-db-sync" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.741757 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eb055cf-7e1b-441b-a1db-4e6479bc9152" containerName="nova-cell0-conductor-db-sync" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.742810 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.745736 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-ckxdc" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.746031 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.761734 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.778166 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a76b16a-3d97-45e9-aefe-3de95463d5fd-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2a76b16a-3d97-45e9-aefe-3de95463d5fd\") " pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.778484 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a76b16a-3d97-45e9-aefe-3de95463d5fd-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2a76b16a-3d97-45e9-aefe-3de95463d5fd\") " pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.778645 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bbrn\" (UniqueName: \"kubernetes.io/projected/2a76b16a-3d97-45e9-aefe-3de95463d5fd-kube-api-access-8bbrn\") pod \"nova-cell0-conductor-0\" (UID: \"2a76b16a-3d97-45e9-aefe-3de95463d5fd\") " pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.880652 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a76b16a-3d97-45e9-aefe-3de95463d5fd-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2a76b16a-3d97-45e9-aefe-3de95463d5fd\") " pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.880711 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bbrn\" (UniqueName: \"kubernetes.io/projected/2a76b16a-3d97-45e9-aefe-3de95463d5fd-kube-api-access-8bbrn\") pod \"nova-cell0-conductor-0\" (UID: \"2a76b16a-3d97-45e9-aefe-3de95463d5fd\") " pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.880876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a76b16a-3d97-45e9-aefe-3de95463d5fd-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2a76b16a-3d97-45e9-aefe-3de95463d5fd\") " pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.884559 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a76b16a-3d97-45e9-aefe-3de95463d5fd-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2a76b16a-3d97-45e9-aefe-3de95463d5fd\") " pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.885658 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a76b16a-3d97-45e9-aefe-3de95463d5fd-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2a76b16a-3d97-45e9-aefe-3de95463d5fd\") " pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:54 crc kubenswrapper[4744]: I0106 15:01:54.896446 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bbrn\" (UniqueName: \"kubernetes.io/projected/2a76b16a-3d97-45e9-aefe-3de95463d5fd-kube-api-access-8bbrn\") pod \"nova-cell0-conductor-0\" (UID: \"2a76b16a-3d97-45e9-aefe-3de95463d5fd\") " pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:55 crc kubenswrapper[4744]: I0106 15:01:55.054860 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:01:55 crc kubenswrapper[4744]: I0106 15:01:55.106959 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:55 crc kubenswrapper[4744]: I0106 15:01:55.167393 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerStarted","Data":"3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec"} Jan 06 15:01:55 crc kubenswrapper[4744]: W0106 15:01:55.677772 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a76b16a_3d97_45e9_aefe_3de95463d5fd.slice/crio-a3ca52164e2da4fd8e4dbeca17bfefe3a7f25bcba906bfb0ad0e8b66cd1c8f0b WatchSource:0}: Error finding container a3ca52164e2da4fd8e4dbeca17bfefe3a7f25bcba906bfb0ad0e8b66cd1c8f0b: Status 404 returned error can't find the container with id a3ca52164e2da4fd8e4dbeca17bfefe3a7f25bcba906bfb0ad0e8b66cd1c8f0b Jan 06 15:01:55 crc kubenswrapper[4744]: I0106 15:01:55.679818 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 06 15:01:56 crc kubenswrapper[4744]: I0106 15:01:56.180854 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2a76b16a-3d97-45e9-aefe-3de95463d5fd","Type":"ContainerStarted","Data":"a3ca52164e2da4fd8e4dbeca17bfefe3a7f25bcba906bfb0ad0e8b66cd1c8f0b"} Jan 06 15:01:56 crc kubenswrapper[4744]: I0106 15:01:56.183724 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerStarted","Data":"b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb"} Jan 06 15:01:57 crc kubenswrapper[4744]: I0106 15:01:57.201071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerStarted","Data":"9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e"} Jan 06 15:01:57 crc kubenswrapper[4744]: I0106 15:01:57.204389 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2a76b16a-3d97-45e9-aefe-3de95463d5fd","Type":"ContainerStarted","Data":"f0d4259289d8f324583dfd6c6becca3b8f925c456f8f9d931af4035a0e845e2e"} Jan 06 15:01:57 crc kubenswrapper[4744]: I0106 15:01:57.204551 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 06 15:01:57 crc kubenswrapper[4744]: I0106 15:01:57.224411 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.224393396 podStartE2EDuration="3.224393396s" podCreationTimestamp="2026-01-06 15:01:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:01:57.220146783 +0000 UTC m=+1513.847613111" watchObservedRunningTime="2026-01-06 15:01:57.224393396 +0000 UTC m=+1513.851859724" Jan 06 15:01:59 crc kubenswrapper[4744]: I0106 15:01:59.228651 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerStarted","Data":"26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab"} Jan 06 15:01:59 crc kubenswrapper[4744]: I0106 15:01:59.229228 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 15:01:59 crc kubenswrapper[4744]: I0106 15:01:59.228879 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="sg-core" containerID="cri-o://9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e" gracePeriod=30 Jan 06 15:01:59 crc kubenswrapper[4744]: I0106 15:01:59.228830 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="ceilometer-central-agent" containerID="cri-o://3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec" gracePeriod=30 Jan 06 15:01:59 crc kubenswrapper[4744]: I0106 15:01:59.228893 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="proxy-httpd" containerID="cri-o://26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab" gracePeriod=30 Jan 06 15:01:59 crc kubenswrapper[4744]: I0106 15:01:59.228889 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="ceilometer-notification-agent" containerID="cri-o://b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb" gracePeriod=30 Jan 06 15:02:00 crc kubenswrapper[4744]: I0106 15:02:00.244791 4744 generic.go:334] "Generic (PLEG): container finished" podID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerID="26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab" exitCode=0 Jan 06 15:02:00 crc kubenswrapper[4744]: I0106 15:02:00.245121 4744 generic.go:334] "Generic (PLEG): container finished" podID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerID="9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e" exitCode=2 Jan 06 15:02:00 crc kubenswrapper[4744]: I0106 15:02:00.245134 4744 generic.go:334] "Generic (PLEG): container finished" podID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerID="b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb" exitCode=0 Jan 06 15:02:00 crc kubenswrapper[4744]: I0106 15:02:00.244859 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerDied","Data":"26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab"} Jan 06 15:02:00 crc kubenswrapper[4744]: I0106 15:02:00.245195 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerDied","Data":"9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e"} Jan 06 15:02:00 crc kubenswrapper[4744]: I0106 15:02:00.245218 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerDied","Data":"b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb"} Jan 06 15:02:00 crc kubenswrapper[4744]: I0106 15:02:00.710993 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:02:00 crc kubenswrapper[4744]: E0106 15:02:00.711584 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.177809 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.290779 4744 generic.go:334] "Generic (PLEG): container finished" podID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerID="3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec" exitCode=0 Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.290827 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerDied","Data":"3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec"} Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.290859 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a","Type":"ContainerDied","Data":"2165586d9d67fa9714a167052f13c402648746a4b4793f4ae01b6c3344650f17"} Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.290877 4744 scope.go:117] "RemoveContainer" containerID="26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.290901 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.302968 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-combined-ca-bundle\") pod \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.303224 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-run-httpd\") pod \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.303305 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggz5t\" (UniqueName: \"kubernetes.io/projected/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-kube-api-access-ggz5t\") pod \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.303326 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-config-data\") pod \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.303449 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-sg-core-conf-yaml\") pod \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.303485 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-scripts\") pod \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.303524 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-log-httpd\") pod \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\" (UID: \"0317db72-7dd0-4a5f-b1db-b8d79cc9d39a\") " Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.303809 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" (UID: "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.304292 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" (UID: "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.309222 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-scripts" (OuterVolumeSpecName: "scripts") pod "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" (UID: "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.311356 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-kube-api-access-ggz5t" (OuterVolumeSpecName: "kube-api-access-ggz5t") pod "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" (UID: "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a"). InnerVolumeSpecName "kube-api-access-ggz5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.318398 4744 scope.go:117] "RemoveContainer" containerID="9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.349740 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" (UID: "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.407023 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.407057 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggz5t\" (UniqueName: \"kubernetes.io/projected/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-kube-api-access-ggz5t\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.407068 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.407077 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.407086 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.470396 4744 scope.go:117] "RemoveContainer" containerID="b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.494359 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" (UID: "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.495332 4744 scope.go:117] "RemoveContainer" containerID="3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.518665 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.544534 4744 scope.go:117] "RemoveContainer" containerID="26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab" Jan 06 15:02:03 crc kubenswrapper[4744]: E0106 15:02:03.556018 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab\": container with ID starting with 26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab not found: ID does not exist" containerID="26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.556179 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab"} err="failed to get container status \"26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab\": rpc error: code = NotFound desc = could not find container \"26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab\": container with ID starting with 26088e3afa34e3493c99a13594770312ace385cf1d855731fe2de546c55905ab not found: ID does not exist" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.556339 4744 scope.go:117] "RemoveContainer" containerID="9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e" Jan 06 15:02:03 crc kubenswrapper[4744]: E0106 15:02:03.556938 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e\": container with ID starting with 9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e not found: ID does not exist" containerID="9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.557015 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e"} err="failed to get container status \"9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e\": rpc error: code = NotFound desc = could not find container \"9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e\": container with ID starting with 9f091a367411107b6bcb794e214a97e919efa013c2b688c84a2b52977522df8e not found: ID does not exist" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.557060 4744 scope.go:117] "RemoveContainer" containerID="b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb" Jan 06 15:02:03 crc kubenswrapper[4744]: E0106 15:02:03.557454 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb\": container with ID starting with b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb not found: ID does not exist" containerID="b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.557474 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb"} err="failed to get container status \"b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb\": rpc error: code = NotFound desc = could not find container \"b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb\": container with ID starting with b65ba208e16bcfab5eb463f8f3407b46893f2ac523123f2ab8d6150aa5f604bb not found: ID does not exist" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.557507 4744 scope.go:117] "RemoveContainer" containerID="3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec" Jan 06 15:02:03 crc kubenswrapper[4744]: E0106 15:02:03.557790 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec\": container with ID starting with 3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec not found: ID does not exist" containerID="3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.557825 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec"} err="failed to get container status \"3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec\": rpc error: code = NotFound desc = could not find container \"3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec\": container with ID starting with 3a335bc1049eebeac5dd80b9a76b769f8e1efcd04ea6455b5cef7c4065a27eec not found: ID does not exist" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.576826 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-config-data" (OuterVolumeSpecName: "config-data") pod "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" (UID: "0317db72-7dd0-4a5f-b1db-b8d79cc9d39a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.683888 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-pvt7q"] Jan 06 15:02:03 crc kubenswrapper[4744]: E0106 15:02:03.684719 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="sg-core" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.684740 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="sg-core" Jan 06 15:02:03 crc kubenswrapper[4744]: E0106 15:02:03.684755 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="proxy-httpd" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.684763 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="proxy-httpd" Jan 06 15:02:03 crc kubenswrapper[4744]: E0106 15:02:03.684807 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="ceilometer-central-agent" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.684814 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="ceilometer-central-agent" Jan 06 15:02:03 crc kubenswrapper[4744]: E0106 15:02:03.684836 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="ceilometer-notification-agent" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.684842 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="ceilometer-notification-agent" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.685285 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="ceilometer-central-agent" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.685330 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="ceilometer-notification-agent" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.685349 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="proxy-httpd" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.685372 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" containerName="sg-core" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.686461 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.700841 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.799242 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-pvt7q"] Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.804390 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm5wb\" (UniqueName: \"kubernetes.io/projected/7ba311c0-54e2-48cd-892b-035b7e44f2d4-kube-api-access-lm5wb\") pod \"aodh-db-create-pvt7q\" (UID: \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\") " pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.804507 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ba311c0-54e2-48cd-892b-035b7e44f2d4-operator-scripts\") pod \"aodh-db-create-pvt7q\" (UID: \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\") " pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.813298 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-4ed5-account-create-update-6k2k5"] Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.818551 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.821379 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.835177 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-4ed5-account-create-update-6k2k5"] Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.873512 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.885622 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.898813 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.904245 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.906786 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.907078 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.909523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j2rg\" (UniqueName: \"kubernetes.io/projected/f3187d5d-d476-46cd-82ef-c4ce080b23ea-kube-api-access-4j2rg\") pod \"aodh-4ed5-account-create-update-6k2k5\" (UID: \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\") " pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.910197 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm5wb\" (UniqueName: \"kubernetes.io/projected/7ba311c0-54e2-48cd-892b-035b7e44f2d4-kube-api-access-lm5wb\") pod \"aodh-db-create-pvt7q\" (UID: \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\") " pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.910308 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ba311c0-54e2-48cd-892b-035b7e44f2d4-operator-scripts\") pod \"aodh-db-create-pvt7q\" (UID: \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\") " pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.910518 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3187d5d-d476-46cd-82ef-c4ce080b23ea-operator-scripts\") pod \"aodh-4ed5-account-create-update-6k2k5\" (UID: \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\") " pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.911148 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ba311c0-54e2-48cd-892b-035b7e44f2d4-operator-scripts\") pod \"aodh-db-create-pvt7q\" (UID: \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\") " pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.911985 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:03 crc kubenswrapper[4744]: I0106 15:02:03.940590 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm5wb\" (UniqueName: \"kubernetes.io/projected/7ba311c0-54e2-48cd-892b-035b7e44f2d4-kube-api-access-lm5wb\") pod \"aodh-db-create-pvt7q\" (UID: \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\") " pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.012948 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-scripts\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.013012 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3187d5d-d476-46cd-82ef-c4ce080b23ea-operator-scripts\") pod \"aodh-4ed5-account-create-update-6k2k5\" (UID: \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\") " pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.013065 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-config-data\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.013209 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.013234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j2rg\" (UniqueName: \"kubernetes.io/projected/f3187d5d-d476-46cd-82ef-c4ce080b23ea-kube-api-access-4j2rg\") pod \"aodh-4ed5-account-create-update-6k2k5\" (UID: \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\") " pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.013265 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.013291 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqpb7\" (UniqueName: \"kubernetes.io/projected/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-kube-api-access-sqpb7\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.013385 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-log-httpd\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.013413 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-run-httpd\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.013685 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3187d5d-d476-46cd-82ef-c4ce080b23ea-operator-scripts\") pod \"aodh-4ed5-account-create-update-6k2k5\" (UID: \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\") " pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.031808 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j2rg\" (UniqueName: \"kubernetes.io/projected/f3187d5d-d476-46cd-82ef-c4ce080b23ea-kube-api-access-4j2rg\") pod \"aodh-4ed5-account-create-update-6k2k5\" (UID: \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\") " pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.074098 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.115272 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.115326 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.115355 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqpb7\" (UniqueName: \"kubernetes.io/projected/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-kube-api-access-sqpb7\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.115432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-log-httpd\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.115454 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-run-httpd\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.115506 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-scripts\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.115535 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-config-data\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.116584 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-log-httpd\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.119283 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-run-httpd\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.120504 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-config-data\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.121312 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.123032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.128891 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-scripts\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.142996 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqpb7\" (UniqueName: \"kubernetes.io/projected/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-kube-api-access-sqpb7\") pod \"ceilometer-0\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.147603 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.220188 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.615913 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-pvt7q"] Jan 06 15:02:04 crc kubenswrapper[4744]: W0106 15:02:04.862228 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3187d5d_d476_46cd_82ef_c4ce080b23ea.slice/crio-0a50274dca43ce5c3785dd5ffa9f4674bd6958b31d3013c5dad43585dc5e8e54 WatchSource:0}: Error finding container 0a50274dca43ce5c3785dd5ffa9f4674bd6958b31d3013c5dad43585dc5e8e54: Status 404 returned error can't find the container with id 0a50274dca43ce5c3785dd5ffa9f4674bd6958b31d3013c5dad43585dc5e8e54 Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.868584 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:04 crc kubenswrapper[4744]: I0106 15:02:04.901281 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-4ed5-account-create-update-6k2k5"] Jan 06 15:02:05 crc kubenswrapper[4744]: I0106 15:02:05.149199 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Jan 06 15:02:05 crc kubenswrapper[4744]: I0106 15:02:05.347052 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-4ed5-account-create-update-6k2k5" event={"ID":"f3187d5d-d476-46cd-82ef-c4ce080b23ea","Type":"ContainerStarted","Data":"0a50274dca43ce5c3785dd5ffa9f4674bd6958b31d3013c5dad43585dc5e8e54"} Jan 06 15:02:05 crc kubenswrapper[4744]: I0106 15:02:05.349313 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pvt7q" event={"ID":"7ba311c0-54e2-48cd-892b-035b7e44f2d4","Type":"ContainerStarted","Data":"66fd8f803082cfcd3bb4e6a6facb14f496ab25a57925feeaeb6d84ba948e3fd4"} Jan 06 15:02:05 crc kubenswrapper[4744]: I0106 15:02:05.350375 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerStarted","Data":"8af8ada1e97c1215b374859da99fca6761fc5578d2f0e99405416bd16984eaf7"} Jan 06 15:02:05 crc kubenswrapper[4744]: I0106 15:02:05.727261 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0317db72-7dd0-4a5f-b1db-b8d79cc9d39a" path="/var/lib/kubelet/pods/0317db72-7dd0-4a5f-b1db-b8d79cc9d39a/volumes" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.212489 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-j59t8"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.240991 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.258628 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.258665 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.286819 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-config-data\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.287786 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.295526 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7xbw\" (UniqueName: \"kubernetes.io/projected/a102390c-a0ee-4423-91fa-b5e789fd2eac-kube-api-access-g7xbw\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.295875 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-scripts\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.309839 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-j59t8"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.400802 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-scripts\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.403421 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-config-data\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.403627 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.403727 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7xbw\" (UniqueName: \"kubernetes.io/projected/a102390c-a0ee-4423-91fa-b5e789fd2eac-kube-api-access-g7xbw\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.408991 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.411338 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.427424 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.433696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pvt7q" event={"ID":"7ba311c0-54e2-48cd-892b-035b7e44f2d4","Type":"ContainerStarted","Data":"83f8c07d400a3b6b77f900ef57d64cc4443225d3c862be1834309544e20b77dc"} Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.439950 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7xbw\" (UniqueName: \"kubernetes.io/projected/a102390c-a0ee-4423-91fa-b5e789fd2eac-kube-api-access-g7xbw\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.444819 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.453280 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.455321 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-4ed5-account-create-update-6k2k5" event={"ID":"f3187d5d-d476-46cd-82ef-c4ce080b23ea","Type":"ContainerStarted","Data":"4cd4b0933ac9f7630420095487c9fa00a431eece7059c4011db386c7446f3d54"} Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.455450 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.458320 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.459066 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-config-data\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.474794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-scripts\") pod \"nova-cell0-cell-mapping-j59t8\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.502149 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.508250 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-config-data\") pod \"nova-scheduler-0\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.508408 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.508497 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.508526 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.508662 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4xqd\" (UniqueName: \"kubernetes.io/projected/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-kube-api-access-r4xqd\") pod \"nova-scheduler-0\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.508686 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dgwf\" (UniqueName: \"kubernetes.io/projected/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-kube-api-access-2dgwf\") pod \"nova-cell1-novncproxy-0\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.518241 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.557565 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.561339 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.571944 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.587122 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-pvt7q" podStartSLOduration=3.587095136 podStartE2EDuration="3.587095136s" podCreationTimestamp="2026-01-06 15:02:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:06.510305323 +0000 UTC m=+1523.137771641" watchObservedRunningTime="2026-01-06 15:02:06.587095136 +0000 UTC m=+1523.214561454" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.613108 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.615012 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.615147 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chxgv\" (UniqueName: \"kubernetes.io/projected/a7b23f9a-6327-401c-a470-69f5d0365632-kube-api-access-chxgv\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.615244 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4xqd\" (UniqueName: \"kubernetes.io/projected/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-kube-api-access-r4xqd\") pod \"nova-scheduler-0\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.615273 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dgwf\" (UniqueName: \"kubernetes.io/projected/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-kube-api-access-2dgwf\") pod \"nova-cell1-novncproxy-0\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.615358 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-config-data\") pod \"nova-scheduler-0\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.615478 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-config-data\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.615527 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.618731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b23f9a-6327-401c-a470-69f5d0365632-logs\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.618861 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.619005 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.639761 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.639901 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-config-data\") pod \"nova-scheduler-0\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.639916 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.640525 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.650079 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.671214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dgwf\" (UniqueName: \"kubernetes.io/projected/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-kube-api-access-2dgwf\") pod \"nova-cell1-novncproxy-0\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.673463 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.675386 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.694271 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.699468 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-4ed5-account-create-update-6k2k5" podStartSLOduration=3.69944332 podStartE2EDuration="3.69944332s" podCreationTimestamp="2026-01-06 15:02:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:06.650144455 +0000 UTC m=+1523.277610773" watchObservedRunningTime="2026-01-06 15:02:06.69944332 +0000 UTC m=+1523.326909638" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.699788 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.708276 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4xqd\" (UniqueName: \"kubernetes.io/projected/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-kube-api-access-r4xqd\") pod \"nova-scheduler-0\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.752970 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfmh6\" (UniqueName: \"kubernetes.io/projected/91f538a4-66fc-4d0b-90af-7ff401447b77-kube-api-access-zfmh6\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.753355 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-config-data\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.753662 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b23f9a-6327-401c-a470-69f5d0365632-logs\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.754007 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91f538a4-66fc-4d0b-90af-7ff401447b77-logs\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.754221 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.759616 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chxgv\" (UniqueName: \"kubernetes.io/projected/a7b23f9a-6327-401c-a470-69f5d0365632-kube-api-access-chxgv\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.759797 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.760040 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-config-data\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.757662 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b23f9a-6327-401c-a470-69f5d0365632-logs\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.766547 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.768678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-config-data\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.826485 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chxgv\" (UniqueName: \"kubernetes.io/projected/a7b23f9a-6327-401c-a470-69f5d0365632-kube-api-access-chxgv\") pod \"nova-api-0\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.864759 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-config-data\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.864955 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfmh6\" (UniqueName: \"kubernetes.io/projected/91f538a4-66fc-4d0b-90af-7ff401447b77-kube-api-access-zfmh6\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.868756 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91f538a4-66fc-4d0b-90af-7ff401447b77-logs\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.869038 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.886568 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91f538a4-66fc-4d0b-90af-7ff401447b77-logs\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.888027 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-config-data\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.898602 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.901021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.906640 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7877d89589-4wt4q"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.909043 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.916905 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.921613 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-4wt4q"] Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.951474 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.957754 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfmh6\" (UniqueName: \"kubernetes.io/projected/91f538a4-66fc-4d0b-90af-7ff401447b77-kube-api-access-zfmh6\") pod \"nova-metadata-0\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " pod="openstack/nova-metadata-0" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.971650 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-config\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.971712 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.971740 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j4v7\" (UniqueName: \"kubernetes.io/projected/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-kube-api-access-2j4v7\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.971856 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-svc\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.971885 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:06 crc kubenswrapper[4744]: I0106 15:02:06.971907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.037960 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.073935 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-svc\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.074121 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.074210 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.074348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-config\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.074460 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.074565 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j4v7\" (UniqueName: \"kubernetes.io/projected/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-kube-api-access-2j4v7\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.075880 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-svc\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.076567 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-nb\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.076934 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-config\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.077296 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-sb\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.081895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-swift-storage-0\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.119236 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j4v7\" (UniqueName: \"kubernetes.io/projected/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-kube-api-access-2j4v7\") pod \"dnsmasq-dns-7877d89589-4wt4q\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.247226 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.474932 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ba311c0-54e2-48cd-892b-035b7e44f2d4" containerID="83f8c07d400a3b6b77f900ef57d64cc4443225d3c862be1834309544e20b77dc" exitCode=0 Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.475228 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pvt7q" event={"ID":"7ba311c0-54e2-48cd-892b-035b7e44f2d4","Type":"ContainerDied","Data":"83f8c07d400a3b6b77f900ef57d64cc4443225d3c862be1834309544e20b77dc"} Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.477197 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerStarted","Data":"aa8a543ca2a0c1b8d2faec870e1528eae33060bf28127e83abb0e8354261f195"} Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.542121 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-j59t8"] Jan 06 15:02:07 crc kubenswrapper[4744]: I0106 15:02:07.781974 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:02:07 crc kubenswrapper[4744]: W0106 15:02:07.856105 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87e70b9c_5e27_4215_8066_2e1aa81e0ed1.slice/crio-dd8df87d3e85c4494dbf74361ed610d12ed2586d04f92441c0c6b2eecca2c213 WatchSource:0}: Error finding container dd8df87d3e85c4494dbf74361ed610d12ed2586d04f92441c0c6b2eecca2c213: Status 404 returned error can't find the container with id dd8df87d3e85c4494dbf74361ed610d12ed2586d04f92441c0c6b2eecca2c213 Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.185243 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-25l7s"] Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.187722 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.191029 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.197373 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.205684 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-25l7s"] Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.213262 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-config-data\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.213320 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-scripts\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.213352 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sp4g\" (UniqueName: \"kubernetes.io/projected/780c730e-33dd-4008-958b-da44ab017991-kube-api-access-6sp4g\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.213378 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.268784 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-4wt4q"] Jan 06 15:02:08 crc kubenswrapper[4744]: W0106 15:02:08.289942 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa57fab9_b04f_4a72_ba2f_2ee0e76e56b2.slice/crio-335a533282b99963b0fcd9f0ab9a9655a39699aee981b6537abbc48307a15a98 WatchSource:0}: Error finding container 335a533282b99963b0fcd9f0ab9a9655a39699aee981b6537abbc48307a15a98: Status 404 returned error can't find the container with id 335a533282b99963b0fcd9f0ab9a9655a39699aee981b6537abbc48307a15a98 Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.292794 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.315813 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-config-data\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.317504 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-scripts\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.317657 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sp4g\" (UniqueName: \"kubernetes.io/projected/780c730e-33dd-4008-958b-da44ab017991-kube-api-access-6sp4g\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.317860 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.327650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.331172 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-scripts\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.338988 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sp4g\" (UniqueName: \"kubernetes.io/projected/780c730e-33dd-4008-958b-da44ab017991-kube-api-access-6sp4g\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.339117 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-config-data\") pod \"nova-cell1-conductor-db-sync-25l7s\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.358751 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.406611 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.509273 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8aea2ebb-3358-4d04-9f92-c4f2e43fb746","Type":"ContainerStarted","Data":"de1d24e90a710f54cbbf84bcb075829ee02ab5b21ee55aab7a1b93878ee0b7c1"} Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.514179 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerStarted","Data":"38a09ba5ab8bcc1a2218a45a6f4bc7b6dd1e0b7d7bf2824f99b6f554b1610679"} Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.519595 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91f538a4-66fc-4d0b-90af-7ff401447b77","Type":"ContainerStarted","Data":"caf3135cbea77a7bb0e6218ec6b6866f3dbdaa9556ba4a286dedb0cdbfd71169"} Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.531261 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.543547 4744 generic.go:334] "Generic (PLEG): container finished" podID="f3187d5d-d476-46cd-82ef-c4ce080b23ea" containerID="4cd4b0933ac9f7630420095487c9fa00a431eece7059c4011db386c7446f3d54" exitCode=0 Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.543638 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-4ed5-account-create-update-6k2k5" event={"ID":"f3187d5d-d476-46cd-82ef-c4ce080b23ea","Type":"ContainerDied","Data":"4cd4b0933ac9f7630420095487c9fa00a431eece7059c4011db386c7446f3d54"} Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.548251 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7b23f9a-6327-401c-a470-69f5d0365632","Type":"ContainerStarted","Data":"f7526e1b623d2274784c247db07b658c850018557f49203dc17b0cbfd940e0c6"} Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.549194 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" event={"ID":"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2","Type":"ContainerStarted","Data":"335a533282b99963b0fcd9f0ab9a9655a39699aee981b6537abbc48307a15a98"} Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.555598 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"87e70b9c-5e27-4215-8066-2e1aa81e0ed1","Type":"ContainerStarted","Data":"dd8df87d3e85c4494dbf74361ed610d12ed2586d04f92441c0c6b2eecca2c213"} Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.580750 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-j59t8" event={"ID":"a102390c-a0ee-4423-91fa-b5e789fd2eac","Type":"ContainerStarted","Data":"904d6f3f7a389ffaf5a5cd0e1fed71c2ff096f7c520bc3dba8c4b628823a5fb1"} Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.580808 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-j59t8" event={"ID":"a102390c-a0ee-4423-91fa-b5e789fd2eac","Type":"ContainerStarted","Data":"2d4d216d52feb5782cf6543cf832ac135992ecec6697f82412c01e3588b51694"} Jan 06 15:02:08 crc kubenswrapper[4744]: I0106 15:02:08.609364 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-j59t8" podStartSLOduration=2.609342303 podStartE2EDuration="2.609342303s" podCreationTimestamp="2026-01-06 15:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:08.597611033 +0000 UTC m=+1525.225077351" watchObservedRunningTime="2026-01-06 15:02:08.609342303 +0000 UTC m=+1525.236808621" Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.193224 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.343454 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-25l7s"] Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.359034 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm5wb\" (UniqueName: \"kubernetes.io/projected/7ba311c0-54e2-48cd-892b-035b7e44f2d4-kube-api-access-lm5wb\") pod \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\" (UID: \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\") " Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.359752 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ba311c0-54e2-48cd-892b-035b7e44f2d4-operator-scripts\") pod \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\" (UID: \"7ba311c0-54e2-48cd-892b-035b7e44f2d4\") " Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.363887 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ba311c0-54e2-48cd-892b-035b7e44f2d4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7ba311c0-54e2-48cd-892b-035b7e44f2d4" (UID: "7ba311c0-54e2-48cd-892b-035b7e44f2d4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.378381 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ba311c0-54e2-48cd-892b-035b7e44f2d4-kube-api-access-lm5wb" (OuterVolumeSpecName: "kube-api-access-lm5wb") pod "7ba311c0-54e2-48cd-892b-035b7e44f2d4" (UID: "7ba311c0-54e2-48cd-892b-035b7e44f2d4"). InnerVolumeSpecName "kube-api-access-lm5wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.471206 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7ba311c0-54e2-48cd-892b-035b7e44f2d4-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.471273 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm5wb\" (UniqueName: \"kubernetes.io/projected/7ba311c0-54e2-48cd-892b-035b7e44f2d4-kube-api-access-lm5wb\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.623740 4744 generic.go:334] "Generic (PLEG): container finished" podID="aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" containerID="133dcfe7ccf5884e7b70cd789c452081c59268bb258bafdf957ceb6c07b9b6b6" exitCode=0 Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.625415 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" event={"ID":"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2","Type":"ContainerDied","Data":"133dcfe7ccf5884e7b70cd789c452081c59268bb258bafdf957ceb6c07b9b6b6"} Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.637513 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-pvt7q" event={"ID":"7ba311c0-54e2-48cd-892b-035b7e44f2d4","Type":"ContainerDied","Data":"66fd8f803082cfcd3bb4e6a6facb14f496ab25a57925feeaeb6d84ba948e3fd4"} Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.637565 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66fd8f803082cfcd3bb4e6a6facb14f496ab25a57925feeaeb6d84ba948e3fd4" Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.637642 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-pvt7q" Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.653211 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerStarted","Data":"b51c5deaf227a2b2a216e09d198b50f74df86861cf09ea1f7f4a4ff574daa328"} Jan 06 15:02:09 crc kubenswrapper[4744]: I0106 15:02:09.662055 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-25l7s" event={"ID":"780c730e-33dd-4008-958b-da44ab017991","Type":"ContainerStarted","Data":"27dd19fe73120f3ee4721012f89dce4174eb810680d8ed90efea3556bf23d9d8"} Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.349302 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.491204 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.500761 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j2rg\" (UniqueName: \"kubernetes.io/projected/f3187d5d-d476-46cd-82ef-c4ce080b23ea-kube-api-access-4j2rg\") pod \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\" (UID: \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\") " Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.500910 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3187d5d-d476-46cd-82ef-c4ce080b23ea-operator-scripts\") pod \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\" (UID: \"f3187d5d-d476-46cd-82ef-c4ce080b23ea\") " Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.502176 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3187d5d-d476-46cd-82ef-c4ce080b23ea-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f3187d5d-d476-46cd-82ef-c4ce080b23ea" (UID: "f3187d5d-d476-46cd-82ef-c4ce080b23ea"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.543363 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3187d5d-d476-46cd-82ef-c4ce080b23ea-kube-api-access-4j2rg" (OuterVolumeSpecName: "kube-api-access-4j2rg") pod "f3187d5d-d476-46cd-82ef-c4ce080b23ea" (UID: "f3187d5d-d476-46cd-82ef-c4ce080b23ea"). InnerVolumeSpecName "kube-api-access-4j2rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.548556 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.604648 4744 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3187d5d-d476-46cd-82ef-c4ce080b23ea-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.604681 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j2rg\" (UniqueName: \"kubernetes.io/projected/f3187d5d-d476-46cd-82ef-c4ce080b23ea-kube-api-access-4j2rg\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.677323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-25l7s" event={"ID":"780c730e-33dd-4008-958b-da44ab017991","Type":"ContainerStarted","Data":"d9413cf78bfc30666af986f6257ca8809ed298701060bc9bc879d08c6a3f6468"} Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.681608 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-4ed5-account-create-update-6k2k5" event={"ID":"f3187d5d-d476-46cd-82ef-c4ce080b23ea","Type":"ContainerDied","Data":"0a50274dca43ce5c3785dd5ffa9f4674bd6958b31d3013c5dad43585dc5e8e54"} Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.681648 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a50274dca43ce5c3785dd5ffa9f4674bd6958b31d3013c5dad43585dc5e8e54" Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.681787 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-4ed5-account-create-update-6k2k5" Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.687321 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" event={"ID":"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2","Type":"ContainerStarted","Data":"e4392110e5cf7cd9c8d6dfb40853a304682005ec5f8a3eb9985b3fe0784ec548"} Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.688139 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.700860 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-25l7s" podStartSLOduration=2.700841572 podStartE2EDuration="2.700841572s" podCreationTimestamp="2026-01-06 15:02:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:10.692682497 +0000 UTC m=+1527.320148815" watchObservedRunningTime="2026-01-06 15:02:10.700841572 +0000 UTC m=+1527.328307890" Jan 06 15:02:10 crc kubenswrapper[4744]: I0106 15:02:10.728529 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" podStartSLOduration=4.728502795 podStartE2EDuration="4.728502795s" podCreationTimestamp="2026-01-06 15:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:10.715374017 +0000 UTC m=+1527.342840345" watchObservedRunningTime="2026-01-06 15:02:10.728502795 +0000 UTC m=+1527.355969113" Jan 06 15:02:11 crc kubenswrapper[4744]: I0106 15:02:11.712372 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:02:11 crc kubenswrapper[4744]: E0106 15:02:11.713736 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:02:11 crc kubenswrapper[4744]: I0106 15:02:11.732740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerStarted","Data":"ec960f329de9914136399c8db920672f7100dfcdca268b09c624d656200e1ba2"} Jan 06 15:02:11 crc kubenswrapper[4744]: I0106 15:02:11.732801 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 15:02:11 crc kubenswrapper[4744]: I0106 15:02:11.760755 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.089998169 podStartE2EDuration="8.760730307s" podCreationTimestamp="2026-01-06 15:02:03 +0000 UTC" firstStartedPulling="2026-01-06 15:02:04.863374741 +0000 UTC m=+1521.490841069" lastFinishedPulling="2026-01-06 15:02:10.534106889 +0000 UTC m=+1527.161573207" observedRunningTime="2026-01-06 15:02:11.748917534 +0000 UTC m=+1528.376383872" watchObservedRunningTime="2026-01-06 15:02:11.760730307 +0000 UTC m=+1528.388196635" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.653973 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-2mm6l"] Jan 06 15:02:14 crc kubenswrapper[4744]: E0106 15:02:14.663043 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ba311c0-54e2-48cd-892b-035b7e44f2d4" containerName="mariadb-database-create" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.663093 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ba311c0-54e2-48cd-892b-035b7e44f2d4" containerName="mariadb-database-create" Jan 06 15:02:14 crc kubenswrapper[4744]: E0106 15:02:14.663144 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3187d5d-d476-46cd-82ef-c4ce080b23ea" containerName="mariadb-account-create-update" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.663152 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3187d5d-d476-46cd-82ef-c4ce080b23ea" containerName="mariadb-account-create-update" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.663835 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3187d5d-d476-46cd-82ef-c4ce080b23ea" containerName="mariadb-account-create-update" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.663859 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ba311c0-54e2-48cd-892b-035b7e44f2d4" containerName="mariadb-database-create" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.665421 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.670718 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.670900 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.679850 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tjbk8" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.680230 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.696923 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-2mm6l"] Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.764244 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8aea2ebb-3358-4d04-9f92-c4f2e43fb746","Type":"ContainerStarted","Data":"c8c44ffb7b70886013bdbf4132702aa0ce06469e2b97d626ca325efe24ac719c"} Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.764417 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="8aea2ebb-3358-4d04-9f92-c4f2e43fb746" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://c8c44ffb7b70886013bdbf4132702aa0ce06469e2b97d626ca325efe24ac719c" gracePeriod=30 Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.772666 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91f538a4-66fc-4d0b-90af-7ff401447b77","Type":"ContainerStarted","Data":"544f6575317f435c90c3406dcde2aefe8831848464bcfe9a2518ebd751bc07c6"} Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.776866 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7b23f9a-6327-401c-a470-69f5d0365632","Type":"ContainerStarted","Data":"d45a2107bad45f86e44c8fc5363ad14d76ce806baf5a1db686e281efd2dae1c5"} Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.784899 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"87e70b9c-5e27-4215-8066-2e1aa81e0ed1","Type":"ContainerStarted","Data":"3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84"} Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.795925 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.732948644 podStartE2EDuration="8.795906585s" podCreationTimestamp="2026-01-06 15:02:06 +0000 UTC" firstStartedPulling="2026-01-06 15:02:08.285993594 +0000 UTC m=+1524.913459912" lastFinishedPulling="2026-01-06 15:02:13.348951535 +0000 UTC m=+1529.976417853" observedRunningTime="2026-01-06 15:02:14.788867368 +0000 UTC m=+1531.416333686" watchObservedRunningTime="2026-01-06 15:02:14.795906585 +0000 UTC m=+1531.423372903" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.816629 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.3274551900000002 podStartE2EDuration="8.816607552s" podCreationTimestamp="2026-01-06 15:02:06 +0000 UTC" firstStartedPulling="2026-01-06 15:02:07.858127709 +0000 UTC m=+1524.485594027" lastFinishedPulling="2026-01-06 15:02:13.347280071 +0000 UTC m=+1529.974746389" observedRunningTime="2026-01-06 15:02:14.80557634 +0000 UTC m=+1531.433042668" watchObservedRunningTime="2026-01-06 15:02:14.816607552 +0000 UTC m=+1531.444073870" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.844535 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj7d4\" (UniqueName: \"kubernetes.io/projected/3361c675-133b-4206-86f1-879f52cdb5bb-kube-api-access-tj7d4\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.844668 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-combined-ca-bundle\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.844736 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-config-data\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.844903 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-scripts\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.946732 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-config-data\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.946876 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-scripts\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.946922 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj7d4\" (UniqueName: \"kubernetes.io/projected/3361c675-133b-4206-86f1-879f52cdb5bb-kube-api-access-tj7d4\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.947096 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-combined-ca-bundle\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.956685 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-config-data\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.957558 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-scripts\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.961878 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-combined-ca-bundle\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:14 crc kubenswrapper[4744]: I0106 15:02:14.980780 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj7d4\" (UniqueName: \"kubernetes.io/projected/3361c675-133b-4206-86f1-879f52cdb5bb-kube-api-access-tj7d4\") pod \"aodh-db-sync-2mm6l\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:15 crc kubenswrapper[4744]: I0106 15:02:15.042096 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:15 crc kubenswrapper[4744]: W0106 15:02:15.696493 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3361c675_133b_4206_86f1_879f52cdb5bb.slice/crio-89a2831129c8d54e3ccb79cffb290dd6915ead2269e37ba3b483823b141f640a WatchSource:0}: Error finding container 89a2831129c8d54e3ccb79cffb290dd6915ead2269e37ba3b483823b141f640a: Status 404 returned error can't find the container with id 89a2831129c8d54e3ccb79cffb290dd6915ead2269e37ba3b483823b141f640a Jan 06 15:02:15 crc kubenswrapper[4744]: I0106 15:02:15.697765 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-2mm6l"] Jan 06 15:02:15 crc kubenswrapper[4744]: I0106 15:02:15.798316 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerName="nova-metadata-log" containerID="cri-o://544f6575317f435c90c3406dcde2aefe8831848464bcfe9a2518ebd751bc07c6" gracePeriod=30 Jan 06 15:02:15 crc kubenswrapper[4744]: I0106 15:02:15.798425 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerName="nova-metadata-metadata" containerID="cri-o://725bbe7373eea5d26aa6c96773e8699674986b1683cddf4694403d71f4d2500d" gracePeriod=30 Jan 06 15:02:15 crc kubenswrapper[4744]: I0106 15:02:15.824453 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.737006681 podStartE2EDuration="9.824433399s" podCreationTimestamp="2026-01-06 15:02:06 +0000 UTC" firstStartedPulling="2026-01-06 15:02:08.266018536 +0000 UTC m=+1524.893484854" lastFinishedPulling="2026-01-06 15:02:13.353445244 +0000 UTC m=+1529.980911572" observedRunningTime="2026-01-06 15:02:15.813590482 +0000 UTC m=+1532.441056800" watchObservedRunningTime="2026-01-06 15:02:15.824433399 +0000 UTC m=+1532.451899717" Jan 06 15:02:15 crc kubenswrapper[4744]: I0106 15:02:15.842847 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.940030135 podStartE2EDuration="9.842829066s" podCreationTimestamp="2026-01-06 15:02:06 +0000 UTC" firstStartedPulling="2026-01-06 15:02:08.446964095 +0000 UTC m=+1525.074430413" lastFinishedPulling="2026-01-06 15:02:13.349763026 +0000 UTC m=+1529.977229344" observedRunningTime="2026-01-06 15:02:15.832144703 +0000 UTC m=+1532.459611031" watchObservedRunningTime="2026-01-06 15:02:15.842829066 +0000 UTC m=+1532.470295384" Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.127048 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91f538a4-66fc-4d0b-90af-7ff401447b77","Type":"ContainerStarted","Data":"725bbe7373eea5d26aa6c96773e8699674986b1683cddf4694403d71f4d2500d"} Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.127516 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7b23f9a-6327-401c-a470-69f5d0365632","Type":"ContainerStarted","Data":"bfbda32710913e7ee8d3e6ce097f00ee3bb669ab8a458bb98f80315331a8144e"} Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.127546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2mm6l" event={"ID":"3361c675-133b-4206-86f1-879f52cdb5bb","Type":"ContainerStarted","Data":"89a2831129c8d54e3ccb79cffb290dd6915ead2269e37ba3b483823b141f640a"} Jan 06 15:02:16 crc kubenswrapper[4744]: E0106 15:02:16.454209 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91f538a4_66fc_4d0b_90af_7ff401447b77.slice/crio-conmon-725bbe7373eea5d26aa6c96773e8699674986b1683cddf4694403d71f4d2500d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91f538a4_66fc_4d0b_90af_7ff401447b77.slice/crio-conmon-544f6575317f435c90c3406dcde2aefe8831848464bcfe9a2518ebd751bc07c6.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.840055 4744 generic.go:334] "Generic (PLEG): container finished" podID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerID="725bbe7373eea5d26aa6c96773e8699674986b1683cddf4694403d71f4d2500d" exitCode=0 Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.840324 4744 generic.go:334] "Generic (PLEG): container finished" podID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerID="544f6575317f435c90c3406dcde2aefe8831848464bcfe9a2518ebd751bc07c6" exitCode=143 Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.842310 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91f538a4-66fc-4d0b-90af-7ff401447b77","Type":"ContainerDied","Data":"725bbe7373eea5d26aa6c96773e8699674986b1683cddf4694403d71f4d2500d"} Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.842336 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91f538a4-66fc-4d0b-90af-7ff401447b77","Type":"ContainerDied","Data":"544f6575317f435c90c3406dcde2aefe8831848464bcfe9a2518ebd751bc07c6"} Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.901574 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.901614 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.917675 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.935541 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.953115 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 06 15:02:16 crc kubenswrapper[4744]: I0106 15:02:16.954412 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.039277 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.039460 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.255577 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.264352 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.425264 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wwj9s"] Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.425845 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" podUID="4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" containerName="dnsmasq-dns" containerID="cri-o://8e0c7fe7358774ebefe9cd8b3b30df82314da210072c2215775129699a862fb4" gracePeriod=10 Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.426015 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfmh6\" (UniqueName: \"kubernetes.io/projected/91f538a4-66fc-4d0b-90af-7ff401447b77-kube-api-access-zfmh6\") pod \"91f538a4-66fc-4d0b-90af-7ff401447b77\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.426128 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-combined-ca-bundle\") pod \"91f538a4-66fc-4d0b-90af-7ff401447b77\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.426333 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-config-data\") pod \"91f538a4-66fc-4d0b-90af-7ff401447b77\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.426368 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91f538a4-66fc-4d0b-90af-7ff401447b77-logs\") pod \"91f538a4-66fc-4d0b-90af-7ff401447b77\" (UID: \"91f538a4-66fc-4d0b-90af-7ff401447b77\") " Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.432790 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91f538a4-66fc-4d0b-90af-7ff401447b77-logs" (OuterVolumeSpecName: "logs") pod "91f538a4-66fc-4d0b-90af-7ff401447b77" (UID: "91f538a4-66fc-4d0b-90af-7ff401447b77"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.446081 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91f538a4-66fc-4d0b-90af-7ff401447b77-kube-api-access-zfmh6" (OuterVolumeSpecName: "kube-api-access-zfmh6") pod "91f538a4-66fc-4d0b-90af-7ff401447b77" (UID: "91f538a4-66fc-4d0b-90af-7ff401447b77"). InnerVolumeSpecName "kube-api-access-zfmh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.520529 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-config-data" (OuterVolumeSpecName: "config-data") pod "91f538a4-66fc-4d0b-90af-7ff401447b77" (UID: "91f538a4-66fc-4d0b-90af-7ff401447b77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.531562 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.531603 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91f538a4-66fc-4d0b-90af-7ff401447b77-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.531618 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfmh6\" (UniqueName: \"kubernetes.io/projected/91f538a4-66fc-4d0b-90af-7ff401447b77-kube-api-access-zfmh6\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.538766 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91f538a4-66fc-4d0b-90af-7ff401447b77" (UID: "91f538a4-66fc-4d0b-90af-7ff401447b77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.633445 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f538a4-66fc-4d0b-90af-7ff401447b77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.885504 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"91f538a4-66fc-4d0b-90af-7ff401447b77","Type":"ContainerDied","Data":"caf3135cbea77a7bb0e6218ec6b6866f3dbdaa9556ba4a286dedb0cdbfd71169"} Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.885577 4744 scope.go:117] "RemoveContainer" containerID="725bbe7373eea5d26aa6c96773e8699674986b1683cddf4694403d71f4d2500d" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.886021 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.934216 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.953289 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.966367 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:17 crc kubenswrapper[4744]: E0106 15:02:17.967183 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerName="nova-metadata-metadata" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.967202 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerName="nova-metadata-metadata" Jan 06 15:02:17 crc kubenswrapper[4744]: E0106 15:02:17.967220 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerName="nova-metadata-log" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.967226 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerName="nova-metadata-log" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.967505 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerName="nova-metadata-log" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.967533 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="91f538a4-66fc-4d0b-90af-7ff401447b77" containerName="nova-metadata-metadata" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.968768 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.973289 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.973659 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.980125 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:17 crc kubenswrapper[4744]: I0106 15:02:17.981845 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.004456 4744 scope.go:117] "RemoveContainer" containerID="544f6575317f435c90c3406dcde2aefe8831848464bcfe9a2518ebd751bc07c6" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.035313 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.244:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.035589 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.244:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.058805 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6752718d-ba08-41d8-b875-03e9882bfc9b-logs\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.058899 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.059056 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-config-data\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.059090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.059117 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnz8f\" (UniqueName: \"kubernetes.io/projected/6752718d-ba08-41d8-b875-03e9882bfc9b-kube-api-access-qnz8f\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.162541 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6752718d-ba08-41d8-b875-03e9882bfc9b-logs\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.162610 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.162670 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-config-data\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.162697 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.162715 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnz8f\" (UniqueName: \"kubernetes.io/projected/6752718d-ba08-41d8-b875-03e9882bfc9b-kube-api-access-qnz8f\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.163579 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6752718d-ba08-41d8-b875-03e9882bfc9b-logs\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.169434 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.188594 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.196915 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-config-data\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.209905 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnz8f\" (UniqueName: \"kubernetes.io/projected/6752718d-ba08-41d8-b875-03e9882bfc9b-kube-api-access-qnz8f\") pod \"nova-metadata-0\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " pod="openstack/nova-metadata-0" Jan 06 15:02:18 crc kubenswrapper[4744]: I0106 15:02:18.300623 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:18.991364 4744 generic.go:334] "Generic (PLEG): container finished" podID="4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" containerID="8e0c7fe7358774ebefe9cd8b3b30df82314da210072c2215775129699a862fb4" exitCode=0 Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:18.993296 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" event={"ID":"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a","Type":"ContainerDied","Data":"8e0c7fe7358774ebefe9cd8b3b30df82314da210072c2215775129699a862fb4"} Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.008196 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.147130 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.245990 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-sb\") pod \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.246079 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-swift-storage-0\") pod \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.246224 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq56h\" (UniqueName: \"kubernetes.io/projected/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-kube-api-access-hq56h\") pod \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.246330 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-svc\") pod \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.246356 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-config\") pod \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.246376 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-nb\") pod \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\" (UID: \"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a\") " Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.267603 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-kube-api-access-hq56h" (OuterVolumeSpecName: "kube-api-access-hq56h") pod "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" (UID: "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a"). InnerVolumeSpecName "kube-api-access-hq56h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.358206 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq56h\" (UniqueName: \"kubernetes.io/projected/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-kube-api-access-hq56h\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.386506 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" (UID: "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.394487 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" (UID: "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.410583 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-config" (OuterVolumeSpecName: "config") pod "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" (UID: "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.433288 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" (UID: "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.464608 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.464654 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.464664 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.464673 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.470985 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" (UID: "4f9050a1-5cef-40e0-bd7b-3eacc6bc729a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.567427 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:19 crc kubenswrapper[4744]: I0106 15:02:19.735072 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91f538a4-66fc-4d0b-90af-7ff401447b77" path="/var/lib/kubelet/pods/91f538a4-66fc-4d0b-90af-7ff401447b77/volumes" Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.016447 4744 generic.go:334] "Generic (PLEG): container finished" podID="a102390c-a0ee-4423-91fa-b5e789fd2eac" containerID="904d6f3f7a389ffaf5a5cd0e1fed71c2ff096f7c520bc3dba8c4b628823a5fb1" exitCode=0 Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.016890 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-j59t8" event={"ID":"a102390c-a0ee-4423-91fa-b5e789fd2eac","Type":"ContainerDied","Data":"904d6f3f7a389ffaf5a5cd0e1fed71c2ff096f7c520bc3dba8c4b628823a5fb1"} Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.021469 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" event={"ID":"4f9050a1-5cef-40e0-bd7b-3eacc6bc729a","Type":"ContainerDied","Data":"5d1bff51a8fbd687e3e0365a6cf19387b2dc6630f066f75fad402b9d88db9180"} Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.021554 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d978555f9-wwj9s" Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.021565 4744 scope.go:117] "RemoveContainer" containerID="8e0c7fe7358774ebefe9cd8b3b30df82314da210072c2215775129699a862fb4" Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.026519 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6752718d-ba08-41d8-b875-03e9882bfc9b","Type":"ContainerStarted","Data":"6f0b52a025bbc7a9eaaf13a57eb76fb52fcb26358957ff3fab02305080cd8428"} Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.026571 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6752718d-ba08-41d8-b875-03e9882bfc9b","Type":"ContainerStarted","Data":"9780ac1d5bc7fb99c9531b105de74b03a0d578d5f8a758bee1c8af8a5f2d022e"} Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.065729 4744 scope.go:117] "RemoveContainer" containerID="1470ffa3a8a033c5e77e522e47d32fdb0464af0fd74f285dd226ee223696052b" Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.070771 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wwj9s"] Jan 06 15:02:20 crc kubenswrapper[4744]: I0106 15:02:20.081220 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d978555f9-wwj9s"] Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.052848 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6752718d-ba08-41d8-b875-03e9882bfc9b","Type":"ContainerStarted","Data":"c5ce9bac7bfb06d7b204a381ee6793e3a03b180dc4276488d6f7c35d2776cf22"} Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.520638 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.618595 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7xbw\" (UniqueName: \"kubernetes.io/projected/a102390c-a0ee-4423-91fa-b5e789fd2eac-kube-api-access-g7xbw\") pod \"a102390c-a0ee-4423-91fa-b5e789fd2eac\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.618756 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-scripts\") pod \"a102390c-a0ee-4423-91fa-b5e789fd2eac\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.618974 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-config-data\") pod \"a102390c-a0ee-4423-91fa-b5e789fd2eac\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.619004 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-combined-ca-bundle\") pod \"a102390c-a0ee-4423-91fa-b5e789fd2eac\" (UID: \"a102390c-a0ee-4423-91fa-b5e789fd2eac\") " Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.626423 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-scripts" (OuterVolumeSpecName: "scripts") pod "a102390c-a0ee-4423-91fa-b5e789fd2eac" (UID: "a102390c-a0ee-4423-91fa-b5e789fd2eac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.633587 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a102390c-a0ee-4423-91fa-b5e789fd2eac-kube-api-access-g7xbw" (OuterVolumeSpecName: "kube-api-access-g7xbw") pod "a102390c-a0ee-4423-91fa-b5e789fd2eac" (UID: "a102390c-a0ee-4423-91fa-b5e789fd2eac"). InnerVolumeSpecName "kube-api-access-g7xbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.660912 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-config-data" (OuterVolumeSpecName: "config-data") pod "a102390c-a0ee-4423-91fa-b5e789fd2eac" (UID: "a102390c-a0ee-4423-91fa-b5e789fd2eac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.675628 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a102390c-a0ee-4423-91fa-b5e789fd2eac" (UID: "a102390c-a0ee-4423-91fa-b5e789fd2eac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.723223 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.723282 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.723305 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7xbw\" (UniqueName: \"kubernetes.io/projected/a102390c-a0ee-4423-91fa-b5e789fd2eac-kube-api-access-g7xbw\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.723323 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a102390c-a0ee-4423-91fa-b5e789fd2eac-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:21 crc kubenswrapper[4744]: I0106 15:02:21.736024 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" path="/var/lib/kubelet/pods/4f9050a1-5cef-40e0-bd7b-3eacc6bc729a/volumes" Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.068693 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-j59t8" event={"ID":"a102390c-a0ee-4423-91fa-b5e789fd2eac","Type":"ContainerDied","Data":"2d4d216d52feb5782cf6543cf832ac135992ecec6697f82412c01e3588b51694"} Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.068726 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-j59t8" Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.068748 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d4d216d52feb5782cf6543cf832ac135992ecec6697f82412c01e3588b51694" Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.114626 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=5.114605123 podStartE2EDuration="5.114605123s" podCreationTimestamp="2026-01-06 15:02:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:22.094216503 +0000 UTC m=+1538.721682871" watchObservedRunningTime="2026-01-06 15:02:22.114605123 +0000 UTC m=+1538.742071441" Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.246892 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.247135 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-log" containerID="cri-o://d45a2107bad45f86e44c8fc5363ad14d76ce806baf5a1db686e281efd2dae1c5" gracePeriod=30 Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.247296 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-api" containerID="cri-o://bfbda32710913e7ee8d3e6ce097f00ee3bb669ab8a458bb98f80315331a8144e" gracePeriod=30 Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.273776 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.274048 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="87e70b9c-5e27-4215-8066-2e1aa81e0ed1" containerName="nova-scheduler-scheduler" containerID="cri-o://3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84" gracePeriod=30 Jan 06 15:02:22 crc kubenswrapper[4744]: I0106 15:02:22.284362 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:23 crc kubenswrapper[4744]: I0106 15:02:23.078428 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerName="nova-metadata-log" containerID="cri-o://6f0b52a025bbc7a9eaaf13a57eb76fb52fcb26358957ff3fab02305080cd8428" gracePeriod=30 Jan 06 15:02:23 crc kubenswrapper[4744]: I0106 15:02:23.078477 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerName="nova-metadata-metadata" containerID="cri-o://c5ce9bac7bfb06d7b204a381ee6793e3a03b180dc4276488d6f7c35d2776cf22" gracePeriod=30 Jan 06 15:02:23 crc kubenswrapper[4744]: I0106 15:02:23.301302 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 06 15:02:23 crc kubenswrapper[4744]: I0106 15:02:23.301371 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.262590 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dzl9f"] Jan 06 15:02:24 crc kubenswrapper[4744]: E0106 15:02:24.263620 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" containerName="dnsmasq-dns" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.263637 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" containerName="dnsmasq-dns" Jan 06 15:02:24 crc kubenswrapper[4744]: E0106 15:02:24.263693 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" containerName="init" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.263701 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" containerName="init" Jan 06 15:02:24 crc kubenswrapper[4744]: E0106 15:02:24.263751 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a102390c-a0ee-4423-91fa-b5e789fd2eac" containerName="nova-manage" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.263758 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a102390c-a0ee-4423-91fa-b5e789fd2eac" containerName="nova-manage" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.264168 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f9050a1-5cef-40e0-bd7b-3eacc6bc729a" containerName="dnsmasq-dns" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.264194 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a102390c-a0ee-4423-91fa-b5e789fd2eac" containerName="nova-manage" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.266979 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.279943 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzl9f"] Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.296440 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-utilities\") pod \"redhat-marketplace-dzl9f\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.296637 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-catalog-content\") pod \"redhat-marketplace-dzl9f\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.296718 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjspm\" (UniqueName: \"kubernetes.io/projected/1890b419-b512-4925-993a-441edb67ba88-kube-api-access-sjspm\") pod \"redhat-marketplace-dzl9f\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.399147 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-utilities\") pod \"redhat-marketplace-dzl9f\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.399292 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-catalog-content\") pod \"redhat-marketplace-dzl9f\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.399355 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjspm\" (UniqueName: \"kubernetes.io/projected/1890b419-b512-4925-993a-441edb67ba88-kube-api-access-sjspm\") pod \"redhat-marketplace-dzl9f\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.399729 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-catalog-content\") pod \"redhat-marketplace-dzl9f\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.399940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-utilities\") pod \"redhat-marketplace-dzl9f\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.426563 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjspm\" (UniqueName: \"kubernetes.io/projected/1890b419-b512-4925-993a-441edb67ba88-kube-api-access-sjspm\") pod \"redhat-marketplace-dzl9f\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.621173 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:24 crc kubenswrapper[4744]: I0106 15:02:24.711828 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:02:24 crc kubenswrapper[4744]: E0106 15:02:24.712262 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:02:25 crc kubenswrapper[4744]: I0106 15:02:25.103625 4744 generic.go:334] "Generic (PLEG): container finished" podID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerID="c5ce9bac7bfb06d7b204a381ee6793e3a03b180dc4276488d6f7c35d2776cf22" exitCode=0 Jan 06 15:02:25 crc kubenswrapper[4744]: I0106 15:02:25.103859 4744 generic.go:334] "Generic (PLEG): container finished" podID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerID="6f0b52a025bbc7a9eaaf13a57eb76fb52fcb26358957ff3fab02305080cd8428" exitCode=143 Jan 06 15:02:25 crc kubenswrapper[4744]: I0106 15:02:25.103687 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6752718d-ba08-41d8-b875-03e9882bfc9b","Type":"ContainerDied","Data":"c5ce9bac7bfb06d7b204a381ee6793e3a03b180dc4276488d6f7c35d2776cf22"} Jan 06 15:02:25 crc kubenswrapper[4744]: I0106 15:02:25.103924 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6752718d-ba08-41d8-b875-03e9882bfc9b","Type":"ContainerDied","Data":"6f0b52a025bbc7a9eaaf13a57eb76fb52fcb26358957ff3fab02305080cd8428"} Jan 06 15:02:25 crc kubenswrapper[4744]: I0106 15:02:25.107247 4744 generic.go:334] "Generic (PLEG): container finished" podID="a7b23f9a-6327-401c-a470-69f5d0365632" containerID="d45a2107bad45f86e44c8fc5363ad14d76ce806baf5a1db686e281efd2dae1c5" exitCode=143 Jan 06 15:02:25 crc kubenswrapper[4744]: I0106 15:02:25.107292 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7b23f9a-6327-401c-a470-69f5d0365632","Type":"ContainerDied","Data":"d45a2107bad45f86e44c8fc5363ad14d76ce806baf5a1db686e281efd2dae1c5"} Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.122125 4744 generic.go:334] "Generic (PLEG): container finished" podID="87e70b9c-5e27-4215-8066-2e1aa81e0ed1" containerID="3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84" exitCode=0 Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.122211 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"87e70b9c-5e27-4215-8066-2e1aa81e0ed1","Type":"ContainerDied","Data":"3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84"} Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.124813 4744 generic.go:334] "Generic (PLEG): container finished" podID="a7b23f9a-6327-401c-a470-69f5d0365632" containerID="bfbda32710913e7ee8d3e6ce097f00ee3bb669ab8a458bb98f80315331a8144e" exitCode=0 Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.124846 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7b23f9a-6327-401c-a470-69f5d0365632","Type":"ContainerDied","Data":"bfbda32710913e7ee8d3e6ce097f00ee3bb669ab8a458bb98f80315331a8144e"} Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.861821 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.863727 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6752718d-ba08-41d8-b875-03e9882bfc9b-logs\") pod \"6752718d-ba08-41d8-b875-03e9882bfc9b\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.863906 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-combined-ca-bundle\") pod \"6752718d-ba08-41d8-b875-03e9882bfc9b\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.863958 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-config-data\") pod \"6752718d-ba08-41d8-b875-03e9882bfc9b\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.864050 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-nova-metadata-tls-certs\") pod \"6752718d-ba08-41d8-b875-03e9882bfc9b\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.864055 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6752718d-ba08-41d8-b875-03e9882bfc9b-logs" (OuterVolumeSpecName: "logs") pod "6752718d-ba08-41d8-b875-03e9882bfc9b" (UID: "6752718d-ba08-41d8-b875-03e9882bfc9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.864126 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnz8f\" (UniqueName: \"kubernetes.io/projected/6752718d-ba08-41d8-b875-03e9882bfc9b-kube-api-access-qnz8f\") pod \"6752718d-ba08-41d8-b875-03e9882bfc9b\" (UID: \"6752718d-ba08-41d8-b875-03e9882bfc9b\") " Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.864642 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6752718d-ba08-41d8-b875-03e9882bfc9b-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.871361 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6752718d-ba08-41d8-b875-03e9882bfc9b-kube-api-access-qnz8f" (OuterVolumeSpecName: "kube-api-access-qnz8f") pod "6752718d-ba08-41d8-b875-03e9882bfc9b" (UID: "6752718d-ba08-41d8-b875-03e9882bfc9b"). InnerVolumeSpecName "kube-api-access-qnz8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:26 crc kubenswrapper[4744]: E0106 15:02:26.902028 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84 is running failed: container process not found" containerID="3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 06 15:02:26 crc kubenswrapper[4744]: E0106 15:02:26.902588 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84 is running failed: container process not found" containerID="3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.902750 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6752718d-ba08-41d8-b875-03e9882bfc9b" (UID: "6752718d-ba08-41d8-b875-03e9882bfc9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:26 crc kubenswrapper[4744]: E0106 15:02:26.903029 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84 is running failed: container process not found" containerID="3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 06 15:02:26 crc kubenswrapper[4744]: E0106 15:02:26.903116 4744 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="87e70b9c-5e27-4215-8066-2e1aa81e0ed1" containerName="nova-scheduler-scheduler" Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.946392 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-config-data" (OuterVolumeSpecName: "config-data") pod "6752718d-ba08-41d8-b875-03e9882bfc9b" (UID: "6752718d-ba08-41d8-b875-03e9882bfc9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.966926 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnz8f\" (UniqueName: \"kubernetes.io/projected/6752718d-ba08-41d8-b875-03e9882bfc9b-kube-api-access-qnz8f\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.966960 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.966972 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:26 crc kubenswrapper[4744]: I0106 15:02:26.967443 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6752718d-ba08-41d8-b875-03e9882bfc9b" (UID: "6752718d-ba08-41d8-b875-03e9882bfc9b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.071264 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6752718d-ba08-41d8-b875-03e9882bfc9b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.140662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6752718d-ba08-41d8-b875-03e9882bfc9b","Type":"ContainerDied","Data":"9780ac1d5bc7fb99c9531b105de74b03a0d578d5f8a758bee1c8af8a5f2d022e"} Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.140708 4744 scope.go:117] "RemoveContainer" containerID="c5ce9bac7bfb06d7b204a381ee6793e3a03b180dc4276488d6f7c35d2776cf22" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.140712 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.186509 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.209381 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.221212 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:27 crc kubenswrapper[4744]: E0106 15:02:27.221880 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerName="nova-metadata-log" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.221903 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerName="nova-metadata-log" Jan 06 15:02:27 crc kubenswrapper[4744]: E0106 15:02:27.221920 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerName="nova-metadata-metadata" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.221927 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerName="nova-metadata-metadata" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.222151 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerName="nova-metadata-log" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.222192 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6752718d-ba08-41d8-b875-03e9882bfc9b" containerName="nova-metadata-metadata" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.223470 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.225999 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.226699 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.234987 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.283132 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.283203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.283226 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9255f27d-3d50-47f7-92b0-ee9331a565cd-logs\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.283339 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t687x\" (UniqueName: \"kubernetes.io/projected/9255f27d-3d50-47f7-92b0-ee9331a565cd-kube-api-access-t687x\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.283408 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-config-data\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.385345 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t687x\" (UniqueName: \"kubernetes.io/projected/9255f27d-3d50-47f7-92b0-ee9331a565cd-kube-api-access-t687x\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.385654 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-config-data\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.385716 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.385747 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.385764 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9255f27d-3d50-47f7-92b0-ee9331a565cd-logs\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.386538 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9255f27d-3d50-47f7-92b0-ee9331a565cd-logs\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.389266 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-config-data\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.390703 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.398269 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.401263 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t687x\" (UniqueName: \"kubernetes.io/projected/9255f27d-3d50-47f7-92b0-ee9331a565cd-kube-api-access-t687x\") pod \"nova-metadata-0\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.583027 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.584214 4744 scope.go:117] "RemoveContainer" containerID="6f0b52a025bbc7a9eaaf13a57eb76fb52fcb26358957ff3fab02305080cd8428" Jan 06 15:02:27 crc kubenswrapper[4744]: I0106 15:02:27.728741 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6752718d-ba08-41d8-b875-03e9882bfc9b" path="/var/lib/kubelet/pods/6752718d-ba08-41d8-b875-03e9882bfc9b/volumes" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.175563 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.176093 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"87e70b9c-5e27-4215-8066-2e1aa81e0ed1","Type":"ContainerDied","Data":"dd8df87d3e85c4494dbf74361ed610d12ed2586d04f92441c0c6b2eecca2c213"} Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.176145 4744 scope.go:117] "RemoveContainer" containerID="3cbd71a5248f8b9b955f53f085085abb8872de3a873d59dd441875cf8d3dcc84" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.184234 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a7b23f9a-6327-401c-a470-69f5d0365632","Type":"ContainerDied","Data":"f7526e1b623d2274784c247db07b658c850018557f49203dc17b0cbfd940e0c6"} Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.184283 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7526e1b623d2274784c247db07b658c850018557f49203dc17b0cbfd940e0c6" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.187248 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.189641 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2mm6l" event={"ID":"3361c675-133b-4206-86f1-879f52cdb5bb","Type":"ContainerStarted","Data":"2a70419d4e1df7b45905a583313565b14f3fc14270df6b3cadc52370b7a9d58b"} Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.225189 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-config-data\") pod \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.225255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chxgv\" (UniqueName: \"kubernetes.io/projected/a7b23f9a-6327-401c-a470-69f5d0365632-kube-api-access-chxgv\") pod \"a7b23f9a-6327-401c-a470-69f5d0365632\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.225347 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b23f9a-6327-401c-a470-69f5d0365632-logs\") pod \"a7b23f9a-6327-401c-a470-69f5d0365632\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.226333 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-combined-ca-bundle\") pod \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.226635 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-combined-ca-bundle\") pod \"a7b23f9a-6327-401c-a470-69f5d0365632\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.226794 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4xqd\" (UniqueName: \"kubernetes.io/projected/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-kube-api-access-r4xqd\") pod \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\" (UID: \"87e70b9c-5e27-4215-8066-2e1aa81e0ed1\") " Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.226838 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-config-data\") pod \"a7b23f9a-6327-401c-a470-69f5d0365632\" (UID: \"a7b23f9a-6327-401c-a470-69f5d0365632\") " Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.230787 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7b23f9a-6327-401c-a470-69f5d0365632-logs" (OuterVolumeSpecName: "logs") pod "a7b23f9a-6327-401c-a470-69f5d0365632" (UID: "a7b23f9a-6327-401c-a470-69f5d0365632"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.242480 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-kube-api-access-r4xqd" (OuterVolumeSpecName: "kube-api-access-r4xqd") pod "87e70b9c-5e27-4215-8066-2e1aa81e0ed1" (UID: "87e70b9c-5e27-4215-8066-2e1aa81e0ed1"). InnerVolumeSpecName "kube-api-access-r4xqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:28 crc kubenswrapper[4744]: W0106 15:02:28.242937 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1890b419_b512_4925_993a_441edb67ba88.slice/crio-16d7a730d44544f2cb3486f60856d3c62a36cbcbe3c671587654e9b39ebe7e18 WatchSource:0}: Error finding container 16d7a730d44544f2cb3486f60856d3c62a36cbcbe3c671587654e9b39ebe7e18: Status 404 returned error can't find the container with id 16d7a730d44544f2cb3486f60856d3c62a36cbcbe3c671587654e9b39ebe7e18 Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.266623 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7b23f9a-6327-401c-a470-69f5d0365632-kube-api-access-chxgv" (OuterVolumeSpecName: "kube-api-access-chxgv") pod "a7b23f9a-6327-401c-a470-69f5d0365632" (UID: "a7b23f9a-6327-401c-a470-69f5d0365632"). InnerVolumeSpecName "kube-api-access-chxgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.305338 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzl9f"] Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.309957 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-config-data" (OuterVolumeSpecName: "config-data") pod "a7b23f9a-6327-401c-a470-69f5d0365632" (UID: "a7b23f9a-6327-401c-a470-69f5d0365632"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.328075 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-2mm6l" podStartSLOduration=2.444232405 podStartE2EDuration="14.328052956s" podCreationTimestamp="2026-01-06 15:02:14 +0000 UTC" firstStartedPulling="2026-01-06 15:02:15.700246752 +0000 UTC m=+1532.327713070" lastFinishedPulling="2026-01-06 15:02:27.584067303 +0000 UTC m=+1544.211533621" observedRunningTime="2026-01-06 15:02:28.261370141 +0000 UTC m=+1544.888836459" watchObservedRunningTime="2026-01-06 15:02:28.328052956 +0000 UTC m=+1544.955519274" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.331084 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7b23f9a-6327-401c-a470-69f5d0365632-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.331114 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4xqd\" (UniqueName: \"kubernetes.io/projected/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-kube-api-access-r4xqd\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.331128 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.331139 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chxgv\" (UniqueName: \"kubernetes.io/projected/a7b23f9a-6327-401c-a470-69f5d0365632-kube-api-access-chxgv\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.369113 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87e70b9c-5e27-4215-8066-2e1aa81e0ed1" (UID: "87e70b9c-5e27-4215-8066-2e1aa81e0ed1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.372326 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.391948 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-config-data" (OuterVolumeSpecName: "config-data") pod "87e70b9c-5e27-4215-8066-2e1aa81e0ed1" (UID: "87e70b9c-5e27-4215-8066-2e1aa81e0ed1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.401118 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7b23f9a-6327-401c-a470-69f5d0365632" (UID: "a7b23f9a-6327-401c-a470-69f5d0365632"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.433621 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.433669 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e70b9c-5e27-4215-8066-2e1aa81e0ed1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:28 crc kubenswrapper[4744]: I0106 15:02:28.433685 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7b23f9a-6327-401c-a470-69f5d0365632-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.208524 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.221569 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9255f27d-3d50-47f7-92b0-ee9331a565cd","Type":"ContainerStarted","Data":"a88070ac4ba00f8e117f58818e579e3cc3c8cad215bcc3b69647a07f2884409f"} Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.221626 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9255f27d-3d50-47f7-92b0-ee9331a565cd","Type":"ContainerStarted","Data":"96d4d2f7c18c6829a3a079d4f8c558a32b22a7e809303f87b78955a94a9a568f"} Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.221643 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9255f27d-3d50-47f7-92b0-ee9331a565cd","Type":"ContainerStarted","Data":"3372bc0d9f9018b84e81bab2ae44e78ef49a9a3621023291179918c6a4f14919"} Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.229792 4744 generic.go:334] "Generic (PLEG): container finished" podID="1890b419-b512-4925-993a-441edb67ba88" containerID="1476096185dfdfff11175eb50136ac0dc71dd503fe37b23bd1c30f0d27a95668" exitCode=0 Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.229880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzl9f" event={"ID":"1890b419-b512-4925-993a-441edb67ba88","Type":"ContainerDied","Data":"1476096185dfdfff11175eb50136ac0dc71dd503fe37b23bd1c30f0d27a95668"} Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.229909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzl9f" event={"ID":"1890b419-b512-4925-993a-441edb67ba88","Type":"ContainerStarted","Data":"16d7a730d44544f2cb3486f60856d3c62a36cbcbe3c671587654e9b39ebe7e18"} Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.237732 4744 generic.go:334] "Generic (PLEG): container finished" podID="780c730e-33dd-4008-958b-da44ab017991" containerID="d9413cf78bfc30666af986f6257ca8809ed298701060bc9bc879d08c6a3f6468" exitCode=0 Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.237821 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-25l7s" event={"ID":"780c730e-33dd-4008-958b-da44ab017991","Type":"ContainerDied","Data":"d9413cf78bfc30666af986f6257ca8809ed298701060bc9bc879d08c6a3f6468"} Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.238085 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.267587 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.267563174 podStartE2EDuration="2.267563174s" podCreationTimestamp="2026-01-06 15:02:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:29.244297488 +0000 UTC m=+1545.871763846" watchObservedRunningTime="2026-01-06 15:02:29.267563174 +0000 UTC m=+1545.895029492" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.289326 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.312221 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.361225 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:02:29 crc kubenswrapper[4744]: E0106 15:02:29.361691 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-log" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.361708 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-log" Jan 06 15:02:29 crc kubenswrapper[4744]: E0106 15:02:29.361731 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e70b9c-5e27-4215-8066-2e1aa81e0ed1" containerName="nova-scheduler-scheduler" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.361737 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e70b9c-5e27-4215-8066-2e1aa81e0ed1" containerName="nova-scheduler-scheduler" Jan 06 15:02:29 crc kubenswrapper[4744]: E0106 15:02:29.361775 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-api" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.361781 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-api" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.361974 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-log" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.361995 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" containerName="nova-api-api" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.362013 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e70b9c-5e27-4215-8066-2e1aa81e0ed1" containerName="nova-scheduler-scheduler" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.362764 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.365082 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.386058 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.408292 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.418848 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.429265 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.431069 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.433041 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.446238 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.456462 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcn2p\" (UniqueName: \"kubernetes.io/projected/c37294a5-5d6b-495a-a392-7744df1c0ec9-kube-api-access-xcn2p\") pod \"nova-scheduler-0\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.456566 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-logs\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.456635 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-config-data\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.456677 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-config-data\") pod \"nova-scheduler-0\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.456707 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.456838 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.456923 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7vmk\" (UniqueName: \"kubernetes.io/projected/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-kube-api-access-s7vmk\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.558754 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-config-data\") pod \"nova-scheduler-0\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.559380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.559658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.559862 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7vmk\" (UniqueName: \"kubernetes.io/projected/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-kube-api-access-s7vmk\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.559987 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcn2p\" (UniqueName: \"kubernetes.io/projected/c37294a5-5d6b-495a-a392-7744df1c0ec9-kube-api-access-xcn2p\") pod \"nova-scheduler-0\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.560302 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-logs\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.562810 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-config-data\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.564999 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-logs\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.565748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-config-data\") pod \"nova-scheduler-0\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.568106 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-config-data\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.568748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.579492 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.581906 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcn2p\" (UniqueName: \"kubernetes.io/projected/c37294a5-5d6b-495a-a392-7744df1c0ec9-kube-api-access-xcn2p\") pod \"nova-scheduler-0\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.585123 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7vmk\" (UniqueName: \"kubernetes.io/projected/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-kube-api-access-s7vmk\") pod \"nova-api-0\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.682316 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.746570 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.747237 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87e70b9c-5e27-4215-8066-2e1aa81e0ed1" path="/var/lib/kubelet/pods/87e70b9c-5e27-4215-8066-2e1aa81e0ed1/volumes" Jan 06 15:02:29 crc kubenswrapper[4744]: I0106 15:02:29.747780 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7b23f9a-6327-401c-a470-69f5d0365632" path="/var/lib/kubelet/pods/a7b23f9a-6327-401c-a470-69f5d0365632/volumes" Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.221772 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:02:30 crc kubenswrapper[4744]: W0106 15:02:30.222006 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc37294a5_5d6b_495a_a392_7744df1c0ec9.slice/crio-512fb1f7b14c7fa98c4928e2822ce2cfc62b0640e93d503d1355b51fe382fc95 WatchSource:0}: Error finding container 512fb1f7b14c7fa98c4928e2822ce2cfc62b0640e93d503d1355b51fe382fc95: Status 404 returned error can't find the container with id 512fb1f7b14c7fa98c4928e2822ce2cfc62b0640e93d503d1355b51fe382fc95 Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.251212 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c37294a5-5d6b-495a-a392-7744df1c0ec9","Type":"ContainerStarted","Data":"512fb1f7b14c7fa98c4928e2822ce2cfc62b0640e93d503d1355b51fe382fc95"} Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.320595 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.609252 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.696106 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-config-data\") pod \"780c730e-33dd-4008-958b-da44ab017991\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.696364 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sp4g\" (UniqueName: \"kubernetes.io/projected/780c730e-33dd-4008-958b-da44ab017991-kube-api-access-6sp4g\") pod \"780c730e-33dd-4008-958b-da44ab017991\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.696402 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-scripts\") pod \"780c730e-33dd-4008-958b-da44ab017991\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.696430 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-combined-ca-bundle\") pod \"780c730e-33dd-4008-958b-da44ab017991\" (UID: \"780c730e-33dd-4008-958b-da44ab017991\") " Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.699904 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-scripts" (OuterVolumeSpecName: "scripts") pod "780c730e-33dd-4008-958b-da44ab017991" (UID: "780c730e-33dd-4008-958b-da44ab017991"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.702007 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/780c730e-33dd-4008-958b-da44ab017991-kube-api-access-6sp4g" (OuterVolumeSpecName: "kube-api-access-6sp4g") pod "780c730e-33dd-4008-958b-da44ab017991" (UID: "780c730e-33dd-4008-958b-da44ab017991"). InnerVolumeSpecName "kube-api-access-6sp4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.729358 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "780c730e-33dd-4008-958b-da44ab017991" (UID: "780c730e-33dd-4008-958b-da44ab017991"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.744045 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-config-data" (OuterVolumeSpecName: "config-data") pod "780c730e-33dd-4008-958b-da44ab017991" (UID: "780c730e-33dd-4008-958b-da44ab017991"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.801656 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sp4g\" (UniqueName: \"kubernetes.io/projected/780c730e-33dd-4008-958b-da44ab017991-kube-api-access-6sp4g\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.801695 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.801707 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:30 crc kubenswrapper[4744]: I0106 15:02:30.801717 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/780c730e-33dd-4008-958b-da44ab017991-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.262579 4744 generic.go:334] "Generic (PLEG): container finished" podID="3361c675-133b-4206-86f1-879f52cdb5bb" containerID="2a70419d4e1df7b45905a583313565b14f3fc14270df6b3cadc52370b7a9d58b" exitCode=0 Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.262668 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2mm6l" event={"ID":"3361c675-133b-4206-86f1-879f52cdb5bb","Type":"ContainerDied","Data":"2a70419d4e1df7b45905a583313565b14f3fc14270df6b3cadc52370b7a9d58b"} Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.264445 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c37294a5-5d6b-495a-a392-7744df1c0ec9","Type":"ContainerStarted","Data":"3d1a9b23a5439d8d400ab9bc66ce612d21b70c49b9b82e16f6d214f81fe399b8"} Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.267472 4744 generic.go:334] "Generic (PLEG): container finished" podID="1890b419-b512-4925-993a-441edb67ba88" containerID="ff1aefd9dc24a0d1d50161f3670fd225a0b0ad283c371078a404122060e11101" exitCode=0 Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.267518 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzl9f" event={"ID":"1890b419-b512-4925-993a-441edb67ba88","Type":"ContainerDied","Data":"ff1aefd9dc24a0d1d50161f3670fd225a0b0ad283c371078a404122060e11101"} Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.271124 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"687c5ae1-31e8-4b4d-9f69-25f3095b10b1","Type":"ContainerStarted","Data":"98872de1df1609af11f2483bc9ef2043eb79b0e9b7b4037a7df854e2a9428182"} Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.271169 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"687c5ae1-31e8-4b4d-9f69-25f3095b10b1","Type":"ContainerStarted","Data":"cf11091fad049ead87f552949b247d2bb1d4a6b4902f4a64b34c79b25ea19bca"} Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.271183 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"687c5ae1-31e8-4b4d-9f69-25f3095b10b1","Type":"ContainerStarted","Data":"37a41e0aef39f1305cb3fb39bc82468053bb560d84814f15c2ceae58e34c7e5c"} Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.275498 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-25l7s" event={"ID":"780c730e-33dd-4008-958b-da44ab017991","Type":"ContainerDied","Data":"27dd19fe73120f3ee4721012f89dce4174eb810680d8ed90efea3556bf23d9d8"} Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.275528 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27dd19fe73120f3ee4721012f89dce4174eb810680d8ed90efea3556bf23d9d8" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.275573 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-25l7s" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.373298 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 06 15:02:31 crc kubenswrapper[4744]: E0106 15:02:31.373904 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="780c730e-33dd-4008-958b-da44ab017991" containerName="nova-cell1-conductor-db-sync" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.373924 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="780c730e-33dd-4008-958b-da44ab017991" containerName="nova-cell1-conductor-db-sync" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.374139 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="780c730e-33dd-4008-958b-da44ab017991" containerName="nova-cell1-conductor-db-sync" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.375052 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.388549 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.388520653 podStartE2EDuration="2.388520653s" podCreationTimestamp="2026-01-06 15:02:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:31.355384036 +0000 UTC m=+1547.982850364" watchObservedRunningTime="2026-01-06 15:02:31.388520653 +0000 UTC m=+1548.015986971" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.388795 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.414291 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e274934-d185-4e44-b1a1-46d78395174c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0e274934-d185-4e44-b1a1-46d78395174c\") " pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.414409 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4k92\" (UniqueName: \"kubernetes.io/projected/0e274934-d185-4e44-b1a1-46d78395174c-kube-api-access-g4k92\") pod \"nova-cell1-conductor-0\" (UID: \"0e274934-d185-4e44-b1a1-46d78395174c\") " pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.414545 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e274934-d185-4e44-b1a1-46d78395174c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0e274934-d185-4e44-b1a1-46d78395174c\") " pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.426233 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.437143 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.437126759 podStartE2EDuration="2.437126759s" podCreationTimestamp="2026-01-06 15:02:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:31.37630655 +0000 UTC m=+1548.003772888" watchObservedRunningTime="2026-01-06 15:02:31.437126759 +0000 UTC m=+1548.064593077" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.516158 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4k92\" (UniqueName: \"kubernetes.io/projected/0e274934-d185-4e44-b1a1-46d78395174c-kube-api-access-g4k92\") pod \"nova-cell1-conductor-0\" (UID: \"0e274934-d185-4e44-b1a1-46d78395174c\") " pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.516274 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e274934-d185-4e44-b1a1-46d78395174c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0e274934-d185-4e44-b1a1-46d78395174c\") " pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.516400 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e274934-d185-4e44-b1a1-46d78395174c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0e274934-d185-4e44-b1a1-46d78395174c\") " pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.524537 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e274934-d185-4e44-b1a1-46d78395174c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0e274934-d185-4e44-b1a1-46d78395174c\") " pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.524549 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e274934-d185-4e44-b1a1-46d78395174c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0e274934-d185-4e44-b1a1-46d78395174c\") " pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.534322 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4k92\" (UniqueName: \"kubernetes.io/projected/0e274934-d185-4e44-b1a1-46d78395174c-kube-api-access-g4k92\") pod \"nova-cell1-conductor-0\" (UID: \"0e274934-d185-4e44-b1a1-46d78395174c\") " pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:31 crc kubenswrapper[4744]: I0106 15:02:31.709299 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:32 crc kubenswrapper[4744]: W0106 15:02:32.245515 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e274934_d185_4e44_b1a1_46d78395174c.slice/crio-6e5f039a1e4f9171199bce6d4e6a43fcafe38c1f4cb39aa52d1478a98acbb38f WatchSource:0}: Error finding container 6e5f039a1e4f9171199bce6d4e6a43fcafe38c1f4cb39aa52d1478a98acbb38f: Status 404 returned error can't find the container with id 6e5f039a1e4f9171199bce6d4e6a43fcafe38c1f4cb39aa52d1478a98acbb38f Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.261618 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.287303 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0e274934-d185-4e44-b1a1-46d78395174c","Type":"ContainerStarted","Data":"6e5f039a1e4f9171199bce6d4e6a43fcafe38c1f4cb39aa52d1478a98acbb38f"} Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.583724 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.583959 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.654092 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.750498 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-scripts\") pod \"3361c675-133b-4206-86f1-879f52cdb5bb\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.750697 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-combined-ca-bundle\") pod \"3361c675-133b-4206-86f1-879f52cdb5bb\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.750735 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-config-data\") pod \"3361c675-133b-4206-86f1-879f52cdb5bb\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.750800 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj7d4\" (UniqueName: \"kubernetes.io/projected/3361c675-133b-4206-86f1-879f52cdb5bb-kube-api-access-tj7d4\") pod \"3361c675-133b-4206-86f1-879f52cdb5bb\" (UID: \"3361c675-133b-4206-86f1-879f52cdb5bb\") " Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.756230 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-scripts" (OuterVolumeSpecName: "scripts") pod "3361c675-133b-4206-86f1-879f52cdb5bb" (UID: "3361c675-133b-4206-86f1-879f52cdb5bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.757290 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3361c675-133b-4206-86f1-879f52cdb5bb-kube-api-access-tj7d4" (OuterVolumeSpecName: "kube-api-access-tj7d4") pod "3361c675-133b-4206-86f1-879f52cdb5bb" (UID: "3361c675-133b-4206-86f1-879f52cdb5bb"). InnerVolumeSpecName "kube-api-access-tj7d4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.784992 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3361c675-133b-4206-86f1-879f52cdb5bb" (UID: "3361c675-133b-4206-86f1-879f52cdb5bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.814332 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-config-data" (OuterVolumeSpecName: "config-data") pod "3361c675-133b-4206-86f1-879f52cdb5bb" (UID: "3361c675-133b-4206-86f1-879f52cdb5bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.854987 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.855019 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.855032 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj7d4\" (UniqueName: \"kubernetes.io/projected/3361c675-133b-4206-86f1-879f52cdb5bb-kube-api-access-tj7d4\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:32 crc kubenswrapper[4744]: I0106 15:02:32.855047 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3361c675-133b-4206-86f1-879f52cdb5bb-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.303726 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0e274934-d185-4e44-b1a1-46d78395174c","Type":"ContainerStarted","Data":"fd66c86d38b9526d4508427b5bbe7b317098cbde93a70edfb67c3f7fa60a9958"} Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.306372 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-2mm6l" event={"ID":"3361c675-133b-4206-86f1-879f52cdb5bb","Type":"ContainerDied","Data":"89a2831129c8d54e3ccb79cffb290dd6915ead2269e37ba3b483823b141f640a"} Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.306415 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89a2831129c8d54e3ccb79cffb290dd6915ead2269e37ba3b483823b141f640a" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.306430 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-2mm6l" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.765960 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Jan 06 15:02:33 crc kubenswrapper[4744]: E0106 15:02:33.766809 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3361c675-133b-4206-86f1-879f52cdb5bb" containerName="aodh-db-sync" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.766823 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3361c675-133b-4206-86f1-879f52cdb5bb" containerName="aodh-db-sync" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.767106 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3361c675-133b-4206-86f1-879f52cdb5bb" containerName="aodh-db-sync" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.783373 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.783499 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.788925 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.789142 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tjbk8" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.789382 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.882371 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-scripts\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.882486 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-config-data\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.882580 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xfw2\" (UniqueName: \"kubernetes.io/projected/70f7184f-9088-49fb-ace5-c14f5efd29e3-kube-api-access-8xfw2\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.882602 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-combined-ca-bundle\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.983962 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-config-data\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.984122 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xfw2\" (UniqueName: \"kubernetes.io/projected/70f7184f-9088-49fb-ace5-c14f5efd29e3-kube-api-access-8xfw2\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.984153 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-combined-ca-bundle\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.984235 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-scripts\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.988895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-scripts\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:33 crc kubenswrapper[4744]: I0106 15:02:33.994751 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-combined-ca-bundle\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:34 crc kubenswrapper[4744]: I0106 15:02:34.003413 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xfw2\" (UniqueName: \"kubernetes.io/projected/70f7184f-9088-49fb-ace5-c14f5efd29e3-kube-api-access-8xfw2\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:34 crc kubenswrapper[4744]: I0106 15:02:34.003916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-config-data\") pod \"aodh-0\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " pod="openstack/aodh-0" Jan 06 15:02:34 crc kubenswrapper[4744]: I0106 15:02:34.151931 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:02:34 crc kubenswrapper[4744]: I0106 15:02:34.228431 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 06 15:02:34 crc kubenswrapper[4744]: I0106 15:02:34.332602 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:34 crc kubenswrapper[4744]: I0106 15:02:34.359056 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.35903322 podStartE2EDuration="3.35903322s" podCreationTimestamp="2026-01-06 15:02:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:34.347656159 +0000 UTC m=+1550.975122477" watchObservedRunningTime="2026-01-06 15:02:34.35903322 +0000 UTC m=+1550.986499538" Jan 06 15:02:34 crc kubenswrapper[4744]: I0106 15:02:34.682952 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 06 15:02:34 crc kubenswrapper[4744]: W0106 15:02:34.732009 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70f7184f_9088_49fb_ace5_c14f5efd29e3.slice/crio-2759e0ed674bc145bcbdd25fda166d06f6d359f160afa69b14cc4f01e732118c WatchSource:0}: Error finding container 2759e0ed674bc145bcbdd25fda166d06f6d359f160afa69b14cc4f01e732118c: Status 404 returned error can't find the container with id 2759e0ed674bc145bcbdd25fda166d06f6d359f160afa69b14cc4f01e732118c Jan 06 15:02:34 crc kubenswrapper[4744]: I0106 15:02:34.736234 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Jan 06 15:02:35 crc kubenswrapper[4744]: I0106 15:02:35.345363 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzl9f" event={"ID":"1890b419-b512-4925-993a-441edb67ba88","Type":"ContainerStarted","Data":"b572173ff58b58b5eadfc7cf6f7b2b1acedf7da7d1ffeb14a2dcc7e98d2db64b"} Jan 06 15:02:35 crc kubenswrapper[4744]: I0106 15:02:35.347527 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerStarted","Data":"2759e0ed674bc145bcbdd25fda166d06f6d359f160afa69b14cc4f01e732118c"} Jan 06 15:02:35 crc kubenswrapper[4744]: I0106 15:02:35.369740 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dzl9f" podStartSLOduration=6.267350297 podStartE2EDuration="11.369715841s" podCreationTimestamp="2026-01-06 15:02:24 +0000 UTC" firstStartedPulling="2026-01-06 15:02:29.231709415 +0000 UTC m=+1545.859175733" lastFinishedPulling="2026-01-06 15:02:34.334074959 +0000 UTC m=+1550.961541277" observedRunningTime="2026-01-06 15:02:35.362030788 +0000 UTC m=+1551.989497106" watchObservedRunningTime="2026-01-06 15:02:35.369715841 +0000 UTC m=+1551.997182169" Jan 06 15:02:36 crc kubenswrapper[4744]: I0106 15:02:36.369877 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerStarted","Data":"8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5"} Jan 06 15:02:36 crc kubenswrapper[4744]: I0106 15:02:36.516610 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:36 crc kubenswrapper[4744]: I0106 15:02:36.517210 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="ceilometer-central-agent" containerID="cri-o://aa8a543ca2a0c1b8d2faec870e1528eae33060bf28127e83abb0e8354261f195" gracePeriod=30 Jan 06 15:02:36 crc kubenswrapper[4744]: I0106 15:02:36.517269 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="proxy-httpd" containerID="cri-o://ec960f329de9914136399c8db920672f7100dfcdca268b09c624d656200e1ba2" gracePeriod=30 Jan 06 15:02:36 crc kubenswrapper[4744]: I0106 15:02:36.517277 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="sg-core" containerID="cri-o://b51c5deaf227a2b2a216e09d198b50f74df86861cf09ea1f7f4a4ff574daa328" gracePeriod=30 Jan 06 15:02:36 crc kubenswrapper[4744]: I0106 15:02:36.517268 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="ceilometer-notification-agent" containerID="cri-o://38a09ba5ab8bcc1a2218a45a6f4bc7b6dd1e0b7d7bf2824f99b6f554b1610679" gracePeriod=30 Jan 06 15:02:36 crc kubenswrapper[4744]: I0106 15:02:36.775175 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Jan 06 15:02:37 crc kubenswrapper[4744]: I0106 15:02:37.584718 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 06 15:02:37 crc kubenswrapper[4744]: I0106 15:02:37.587669 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 06 15:02:38 crc kubenswrapper[4744]: I0106 15:02:38.596320 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.251:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:02:38 crc kubenswrapper[4744]: I0106 15:02:38.596341 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.251:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:02:38 crc kubenswrapper[4744]: I0106 15:02:38.712625 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:02:38 crc kubenswrapper[4744]: E0106 15:02:38.712962 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.441623 4744 generic.go:334] "Generic (PLEG): container finished" podID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerID="ec960f329de9914136399c8db920672f7100dfcdca268b09c624d656200e1ba2" exitCode=0 Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.441858 4744 generic.go:334] "Generic (PLEG): container finished" podID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerID="b51c5deaf227a2b2a216e09d198b50f74df86861cf09ea1f7f4a4ff574daa328" exitCode=2 Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.441868 4744 generic.go:334] "Generic (PLEG): container finished" podID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerID="38a09ba5ab8bcc1a2218a45a6f4bc7b6dd1e0b7d7bf2824f99b6f554b1610679" exitCode=0 Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.441876 4744 generic.go:334] "Generic (PLEG): container finished" podID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerID="aa8a543ca2a0c1b8d2faec870e1528eae33060bf28127e83abb0e8354261f195" exitCode=0 Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.441750 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerDied","Data":"ec960f329de9914136399c8db920672f7100dfcdca268b09c624d656200e1ba2"} Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.441907 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerDied","Data":"b51c5deaf227a2b2a216e09d198b50f74df86861cf09ea1f7f4a4ff574daa328"} Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.441922 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerDied","Data":"38a09ba5ab8bcc1a2218a45a6f4bc7b6dd1e0b7d7bf2824f99b6f554b1610679"} Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.441932 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerDied","Data":"aa8a543ca2a0c1b8d2faec870e1528eae33060bf28127e83abb0e8354261f195"} Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.683930 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.726183 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.747950 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 06 15:02:39 crc kubenswrapper[4744]: I0106 15:02:39.748005 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 06 15:02:40 crc kubenswrapper[4744]: I0106 15:02:40.504839 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 06 15:02:40 crc kubenswrapper[4744]: I0106 15:02:40.833520 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.253:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:02:40 crc kubenswrapper[4744]: I0106 15:02:40.833590 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.253:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.064201 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.180649 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqpb7\" (UniqueName: \"kubernetes.io/projected/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-kube-api-access-sqpb7\") pod \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.181020 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-run-httpd\") pod \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.181125 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-sg-core-conf-yaml\") pod \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.181240 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-config-data\") pod \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.181288 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-scripts\") pod \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.181359 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a83b5a0f-dff5-47ec-be2e-c3788873c9d3" (UID: "a83b5a0f-dff5-47ec-be2e-c3788873c9d3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.181392 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-log-httpd\") pod \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.181558 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-combined-ca-bundle\") pod \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\" (UID: \"a83b5a0f-dff5-47ec-be2e-c3788873c9d3\") " Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.181657 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a83b5a0f-dff5-47ec-be2e-c3788873c9d3" (UID: "a83b5a0f-dff5-47ec-be2e-c3788873c9d3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.183060 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.183086 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.194530 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-kube-api-access-sqpb7" (OuterVolumeSpecName: "kube-api-access-sqpb7") pod "a83b5a0f-dff5-47ec-be2e-c3788873c9d3" (UID: "a83b5a0f-dff5-47ec-be2e-c3788873c9d3"). InnerVolumeSpecName "kube-api-access-sqpb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.216262 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-scripts" (OuterVolumeSpecName: "scripts") pod "a83b5a0f-dff5-47ec-be2e-c3788873c9d3" (UID: "a83b5a0f-dff5-47ec-be2e-c3788873c9d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.256816 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a83b5a0f-dff5-47ec-be2e-c3788873c9d3" (UID: "a83b5a0f-dff5-47ec-be2e-c3788873c9d3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.287770 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqpb7\" (UniqueName: \"kubernetes.io/projected/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-kube-api-access-sqpb7\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.287835 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.287845 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.337406 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a83b5a0f-dff5-47ec-be2e-c3788873c9d3" (UID: "a83b5a0f-dff5-47ec-be2e-c3788873c9d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.389419 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.406194 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-config-data" (OuterVolumeSpecName: "config-data") pod "a83b5a0f-dff5-47ec-be2e-c3788873c9d3" (UID: "a83b5a0f-dff5-47ec-be2e-c3788873c9d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.484555 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.484552 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a83b5a0f-dff5-47ec-be2e-c3788873c9d3","Type":"ContainerDied","Data":"8af8ada1e97c1215b374859da99fca6761fc5578d2f0e99405416bd16984eaf7"} Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.484718 4744 scope.go:117] "RemoveContainer" containerID="ec960f329de9914136399c8db920672f7100dfcdca268b09c624d656200e1ba2" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.496600 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a83b5a0f-dff5-47ec-be2e-c3788873c9d3-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.532340 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.553864 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.569954 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:41 crc kubenswrapper[4744]: E0106 15:02:41.570540 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="ceilometer-central-agent" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.570560 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="ceilometer-central-agent" Jan 06 15:02:41 crc kubenswrapper[4744]: E0106 15:02:41.570577 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="proxy-httpd" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.570584 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="proxy-httpd" Jan 06 15:02:41 crc kubenswrapper[4744]: E0106 15:02:41.570604 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="ceilometer-notification-agent" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.570612 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="ceilometer-notification-agent" Jan 06 15:02:41 crc kubenswrapper[4744]: E0106 15:02:41.570631 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="sg-core" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.570636 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="sg-core" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.570834 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="sg-core" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.570856 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="proxy-httpd" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.570869 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="ceilometer-central-agent" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.570880 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" containerName="ceilometer-notification-agent" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.573010 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.580048 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.583605 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.585193 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.606751 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-config-data\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.606808 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.606901 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-log-httpd\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.606943 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxpzp\" (UniqueName: \"kubernetes.io/projected/89850656-f2cd-413f-9a79-9b08f80c129f-kube-api-access-bxpzp\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.607021 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.607191 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-scripts\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.607337 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-run-httpd\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.709060 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-log-httpd\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.709104 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxpzp\" (UniqueName: \"kubernetes.io/projected/89850656-f2cd-413f-9a79-9b08f80c129f-kube-api-access-bxpzp\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.709132 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.709195 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-scripts\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.709255 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-run-httpd\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.709348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-config-data\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.709401 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.709650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-log-httpd\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.709781 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-run-httpd\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.713764 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-config-data\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.713845 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-scripts\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.717376 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.727098 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxpzp\" (UniqueName: \"kubernetes.io/projected/89850656-f2cd-413f-9a79-9b08f80c129f-kube-api-access-bxpzp\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.734031 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.740959 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a83b5a0f-dff5-47ec-be2e-c3788873c9d3" path="/var/lib/kubelet/pods/a83b5a0f-dff5-47ec-be2e-c3788873c9d3/volumes" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.752743 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.753722 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.776320 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.792466 4744 scope.go:117] "RemoveContainer" containerID="b51c5deaf227a2b2a216e09d198b50f74df86861cf09ea1f7f4a4ff574daa328" Jan 06 15:02:41 crc kubenswrapper[4744]: I0106 15:02:41.961447 4744 scope.go:117] "RemoveContainer" containerID="38a09ba5ab8bcc1a2218a45a6f4bc7b6dd1e0b7d7bf2824f99b6f554b1610679" Jan 06 15:02:42 crc kubenswrapper[4744]: I0106 15:02:42.267751 4744 scope.go:117] "RemoveContainer" containerID="aa8a543ca2a0c1b8d2faec870e1528eae33060bf28127e83abb0e8354261f195" Jan 06 15:02:42 crc kubenswrapper[4744]: W0106 15:02:42.654304 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89850656_f2cd_413f_9a79_9b08f80c129f.slice/crio-5a9cbcc5d5ad34b1816c9fe3d1941cfd57b5074ef5d36f477342cc74028c2e79 WatchSource:0}: Error finding container 5a9cbcc5d5ad34b1816c9fe3d1941cfd57b5074ef5d36f477342cc74028c2e79: Status 404 returned error can't find the container with id 5a9cbcc5d5ad34b1816c9fe3d1941cfd57b5074ef5d36f477342cc74028c2e79 Jan 06 15:02:42 crc kubenswrapper[4744]: I0106 15:02:42.678594 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:02:43 crc kubenswrapper[4744]: I0106 15:02:43.508746 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerStarted","Data":"5a9cbcc5d5ad34b1816c9fe3d1941cfd57b5074ef5d36f477342cc74028c2e79"} Jan 06 15:02:44 crc kubenswrapper[4744]: I0106 15:02:44.622311 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:44 crc kubenswrapper[4744]: I0106 15:02:44.622627 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:44 crc kubenswrapper[4744]: I0106 15:02:44.675470 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:45 crc kubenswrapper[4744]: I0106 15:02:45.585080 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:45 crc kubenswrapper[4744]: I0106 15:02:45.642669 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzl9f"] Jan 06 15:02:46 crc kubenswrapper[4744]: I0106 15:02:46.333292 4744 patch_prober.go:28] interesting pod/route-controller-manager-64bb695654-t9m84 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 15:02:46 crc kubenswrapper[4744]: I0106 15:02:46.333747 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" podUID="8ed4bc19-0cbb-41fe-9956-5aef0d4c287b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 06 15:02:47 crc kubenswrapper[4744]: I0106 15:02:47.280427 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-5649998c8c-g454h" podUID="d18584e4-6488-4fc3-9992-9ded578fd05e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:02:47 crc kubenswrapper[4744]: I0106 15:02:47.552353 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dzl9f" podUID="1890b419-b512-4925-993a-441edb67ba88" containerName="registry-server" containerID="cri-o://b572173ff58b58b5eadfc7cf6f7b2b1acedf7da7d1ffeb14a2dcc7e98d2db64b" gracePeriod=2 Jan 06 15:02:47 crc kubenswrapper[4744]: I0106 15:02:47.590740 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 06 15:02:47 crc kubenswrapper[4744]: I0106 15:02:47.592810 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 06 15:02:47 crc kubenswrapper[4744]: I0106 15:02:47.598774 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 06 15:02:48 crc kubenswrapper[4744]: I0106 15:02:48.574311 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 06 15:02:48 crc kubenswrapper[4744]: I0106 15:02:48.713136 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 06 15:02:48 crc kubenswrapper[4744]: I0106 15:02:48.713345 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="d0377d1a-719f-4733-bbd6-9cd135c1e764" containerName="kube-state-metrics" containerID="cri-o://50186dd7a4c6f1dd389ce2ccd5ea4f266bf4859e95c81c53042661c37d6ca91e" gracePeriod=30 Jan 06 15:02:48 crc kubenswrapper[4744]: I0106 15:02:48.801211 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Jan 06 15:02:48 crc kubenswrapper[4744]: I0106 15:02:48.801446 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="fe94fea2-a713-4982-8d58-9bd7176d99ed" containerName="mysqld-exporter" containerID="cri-o://19395dd7071ace70944a6c11b1d8ecdf40f737a05160e701afe0a94fdc68ea90" gracePeriod=30 Jan 06 15:02:49 crc kubenswrapper[4744]: I0106 15:02:49.588004 4744 generic.go:334] "Generic (PLEG): container finished" podID="d0377d1a-719f-4733-bbd6-9cd135c1e764" containerID="50186dd7a4c6f1dd389ce2ccd5ea4f266bf4859e95c81c53042661c37d6ca91e" exitCode=2 Jan 06 15:02:49 crc kubenswrapper[4744]: I0106 15:02:49.588091 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d0377d1a-719f-4733-bbd6-9cd135c1e764","Type":"ContainerDied","Data":"50186dd7a4c6f1dd389ce2ccd5ea4f266bf4859e95c81c53042661c37d6ca91e"} Jan 06 15:02:49 crc kubenswrapper[4744]: I0106 15:02:49.591656 4744 generic.go:334] "Generic (PLEG): container finished" podID="8aea2ebb-3358-4d04-9f92-c4f2e43fb746" containerID="c8c44ffb7b70886013bdbf4132702aa0ce06469e2b97d626ca325efe24ac719c" exitCode=137 Jan 06 15:02:49 crc kubenswrapper[4744]: I0106 15:02:49.591720 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8aea2ebb-3358-4d04-9f92-c4f2e43fb746","Type":"ContainerDied","Data":"c8c44ffb7b70886013bdbf4132702aa0ce06469e2b97d626ca325efe24ac719c"} Jan 06 15:02:49 crc kubenswrapper[4744]: I0106 15:02:49.753177 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 06 15:02:49 crc kubenswrapper[4744]: I0106 15:02:49.755463 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 06 15:02:49 crc kubenswrapper[4744]: I0106 15:02:49.755515 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 06 15:02:49 crc kubenswrapper[4744]: I0106 15:02:49.757596 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.633856 4744 generic.go:334] "Generic (PLEG): container finished" podID="1890b419-b512-4925-993a-441edb67ba88" containerID="b572173ff58b58b5eadfc7cf6f7b2b1acedf7da7d1ffeb14a2dcc7e98d2db64b" exitCode=0 Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.634346 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzl9f" event={"ID":"1890b419-b512-4925-993a-441edb67ba88","Type":"ContainerDied","Data":"b572173ff58b58b5eadfc7cf6f7b2b1acedf7da7d1ffeb14a2dcc7e98d2db64b"} Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.641406 4744 generic.go:334] "Generic (PLEG): container finished" podID="fe94fea2-a713-4982-8d58-9bd7176d99ed" containerID="19395dd7071ace70944a6c11b1d8ecdf40f737a05160e701afe0a94fdc68ea90" exitCode=2 Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.641499 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"fe94fea2-a713-4982-8d58-9bd7176d99ed","Type":"ContainerDied","Data":"19395dd7071ace70944a6c11b1d8ecdf40f737a05160e701afe0a94fdc68ea90"} Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.652975 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerStarted","Data":"e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5"} Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.653815 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.681194 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.882964 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-x9rf9"] Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.885311 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.897205 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.930471 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-x9rf9"] Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.986606 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnljm\" (UniqueName: \"kubernetes.io/projected/d0377d1a-719f-4733-bbd6-9cd135c1e764-kube-api-access-xnljm\") pod \"d0377d1a-719f-4733-bbd6-9cd135c1e764\" (UID: \"d0377d1a-719f-4733-bbd6-9cd135c1e764\") " Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.986975 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6lmt\" (UniqueName: \"kubernetes.io/projected/7405beaf-33c5-493b-bba7-83419a434632-kube-api-access-r6lmt\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.987022 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.987174 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.987267 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-config\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.987307 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.987332 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:50 crc kubenswrapper[4744]: I0106 15:02:50.992989 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0377d1a-719f-4733-bbd6-9cd135c1e764-kube-api-access-xnljm" (OuterVolumeSpecName: "kube-api-access-xnljm") pod "d0377d1a-719f-4733-bbd6-9cd135c1e764" (UID: "d0377d1a-719f-4733-bbd6-9cd135c1e764"). InnerVolumeSpecName "kube-api-access-xnljm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.090754 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-config\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.091100 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.091352 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.091537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6lmt\" (UniqueName: \"kubernetes.io/projected/7405beaf-33c5-493b-bba7-83419a434632-kube-api-access-r6lmt\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.091673 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.091916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.092091 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnljm\" (UniqueName: \"kubernetes.io/projected/d0377d1a-719f-4733-bbd6-9cd135c1e764-kube-api-access-xnljm\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.094206 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-sb\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.094553 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-svc\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.095074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-nb\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.095279 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-config\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.096891 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-swift-storage-0\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.142655 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6lmt\" (UniqueName: \"kubernetes.io/projected/7405beaf-33c5-493b-bba7-83419a434632-kube-api-access-r6lmt\") pod \"dnsmasq-dns-6d99f6bc7f-x9rf9\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.195508 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.206261 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.251989 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.300101 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-combined-ca-bundle\") pod \"fe94fea2-a713-4982-8d58-9bd7176d99ed\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.300204 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-config-data\") pod \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.300342 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-combined-ca-bundle\") pod \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.300360 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dgwf\" (UniqueName: \"kubernetes.io/projected/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-kube-api-access-2dgwf\") pod \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\" (UID: \"8aea2ebb-3358-4d04-9f92-c4f2e43fb746\") " Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.300443 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6flw\" (UniqueName: \"kubernetes.io/projected/fe94fea2-a713-4982-8d58-9bd7176d99ed-kube-api-access-b6flw\") pod \"fe94fea2-a713-4982-8d58-9bd7176d99ed\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.300513 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-config-data\") pod \"fe94fea2-a713-4982-8d58-9bd7176d99ed\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.308494 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe94fea2-a713-4982-8d58-9bd7176d99ed-kube-api-access-b6flw" (OuterVolumeSpecName: "kube-api-access-b6flw") pod "fe94fea2-a713-4982-8d58-9bd7176d99ed" (UID: "fe94fea2-a713-4982-8d58-9bd7176d99ed"). InnerVolumeSpecName "kube-api-access-b6flw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.314963 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-kube-api-access-2dgwf" (OuterVolumeSpecName: "kube-api-access-2dgwf") pod "8aea2ebb-3358-4d04-9f92-c4f2e43fb746" (UID: "8aea2ebb-3358-4d04-9f92-c4f2e43fb746"). InnerVolumeSpecName "kube-api-access-2dgwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.385417 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-config-data" (OuterVolumeSpecName: "config-data") pod "8aea2ebb-3358-4d04-9f92-c4f2e43fb746" (UID: "8aea2ebb-3358-4d04-9f92-c4f2e43fb746"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.400463 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8aea2ebb-3358-4d04-9f92-c4f2e43fb746" (UID: "8aea2ebb-3358-4d04-9f92-c4f2e43fb746"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.405170 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.405213 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.405227 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dgwf\" (UniqueName: \"kubernetes.io/projected/8aea2ebb-3358-4d04-9f92-c4f2e43fb746-kube-api-access-2dgwf\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.405238 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6flw\" (UniqueName: \"kubernetes.io/projected/fe94fea2-a713-4982-8d58-9bd7176d99ed-kube-api-access-b6flw\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.429274 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe94fea2-a713-4982-8d58-9bd7176d99ed" (UID: "fe94fea2-a713-4982-8d58-9bd7176d99ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.509441 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-config-data" (OuterVolumeSpecName: "config-data") pod "fe94fea2-a713-4982-8d58-9bd7176d99ed" (UID: "fe94fea2-a713-4982-8d58-9bd7176d99ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.509901 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-config-data\") pod \"fe94fea2-a713-4982-8d58-9bd7176d99ed\" (UID: \"fe94fea2-a713-4982-8d58-9bd7176d99ed\") " Jan 06 15:02:51 crc kubenswrapper[4744]: W0106 15:02:51.510326 4744 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/fe94fea2-a713-4982-8d58-9bd7176d99ed/volumes/kubernetes.io~secret/config-data Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.510352 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-config-data" (OuterVolumeSpecName: "config-data") pod "fe94fea2-a713-4982-8d58-9bd7176d99ed" (UID: "fe94fea2-a713-4982-8d58-9bd7176d99ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.510955 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.510969 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe94fea2-a713-4982-8d58-9bd7176d99ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.533341 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.612568 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-catalog-content\") pod \"1890b419-b512-4925-993a-441edb67ba88\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.613107 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjspm\" (UniqueName: \"kubernetes.io/projected/1890b419-b512-4925-993a-441edb67ba88-kube-api-access-sjspm\") pod \"1890b419-b512-4925-993a-441edb67ba88\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.613282 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-utilities\") pod \"1890b419-b512-4925-993a-441edb67ba88\" (UID: \"1890b419-b512-4925-993a-441edb67ba88\") " Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.615611 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-utilities" (OuterVolumeSpecName: "utilities") pod "1890b419-b512-4925-993a-441edb67ba88" (UID: "1890b419-b512-4925-993a-441edb67ba88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.630437 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1890b419-b512-4925-993a-441edb67ba88-kube-api-access-sjspm" (OuterVolumeSpecName: "kube-api-access-sjspm") pod "1890b419-b512-4925-993a-441edb67ba88" (UID: "1890b419-b512-4925-993a-441edb67ba88"). InnerVolumeSpecName "kube-api-access-sjspm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.645101 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1890b419-b512-4925-993a-441edb67ba88" (UID: "1890b419-b512-4925-993a-441edb67ba88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.688020 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dzl9f" event={"ID":"1890b419-b512-4925-993a-441edb67ba88","Type":"ContainerDied","Data":"16d7a730d44544f2cb3486f60856d3c62a36cbcbe3c671587654e9b39ebe7e18"} Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.688072 4744 scope.go:117] "RemoveContainer" containerID="b572173ff58b58b5eadfc7cf6f7b2b1acedf7da7d1ffeb14a2dcc7e98d2db64b" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.688232 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dzl9f" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.702935 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"fe94fea2-a713-4982-8d58-9bd7176d99ed","Type":"ContainerDied","Data":"f909f832b2a68adb2c1b65b2d9c59c6cb9836d3742878a0a29ecb521d2eccbbe"} Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.703030 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.724368 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.724898 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjspm\" (UniqueName: \"kubernetes.io/projected/1890b419-b512-4925-993a-441edb67ba88-kube-api-access-sjspm\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.724959 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1890b419-b512-4925-993a-441edb67ba88-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.727702 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.728336 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:02:51 crc kubenswrapper[4744]: E0106 15:02:51.728865 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.758751 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.916208 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"d0377d1a-719f-4733-bbd6-9cd135c1e764","Type":"ContainerDied","Data":"fae50407dc8919f66deac594799250e2934ff52c111ad5467423db5677615ea6"} Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.916266 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerStarted","Data":"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a"} Jan 06 15:02:51 crc kubenswrapper[4744]: I0106 15:02:51.916284 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8aea2ebb-3358-4d04-9f92-c4f2e43fb746","Type":"ContainerDied","Data":"de1d24e90a710f54cbbf84bcb075829ee02ab5b21ee55aab7a1b93878ee0b7c1"} Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.012550 4744 scope.go:117] "RemoveContainer" containerID="ff1aefd9dc24a0d1d50161f3670fd225a0b0ad283c371078a404122060e11101" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.041272 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-x9rf9"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.069928 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzl9f"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.088103 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dzl9f"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.361347 4744 scope.go:117] "RemoveContainer" containerID="1476096185dfdfff11175eb50136ac0dc71dd503fe37b23bd1c30f0d27a95668" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.425277 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.427085 4744 scope.go:117] "RemoveContainer" containerID="19395dd7071ace70944a6c11b1d8ecdf40f737a05160e701afe0a94fdc68ea90" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.462501 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.484617 4744 scope.go:117] "RemoveContainer" containerID="50186dd7a4c6f1dd389ce2ccd5ea4f266bf4859e95c81c53042661c37d6ca91e" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.496300 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.528704 4744 scope.go:117] "RemoveContainer" containerID="c8c44ffb7b70886013bdbf4132702aa0ce06469e2b97d626ca325efe24ac719c" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.530444 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.543583 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: E0106 15:02:52.544135 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe94fea2-a713-4982-8d58-9bd7176d99ed" containerName="mysqld-exporter" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544183 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe94fea2-a713-4982-8d58-9bd7176d99ed" containerName="mysqld-exporter" Jan 06 15:02:52 crc kubenswrapper[4744]: E0106 15:02:52.544207 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1890b419-b512-4925-993a-441edb67ba88" containerName="extract-utilities" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544215 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1890b419-b512-4925-993a-441edb67ba88" containerName="extract-utilities" Jan 06 15:02:52 crc kubenswrapper[4744]: E0106 15:02:52.544228 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aea2ebb-3358-4d04-9f92-c4f2e43fb746" containerName="nova-cell1-novncproxy-novncproxy" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544237 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aea2ebb-3358-4d04-9f92-c4f2e43fb746" containerName="nova-cell1-novncproxy-novncproxy" Jan 06 15:02:52 crc kubenswrapper[4744]: E0106 15:02:52.544255 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0377d1a-719f-4733-bbd6-9cd135c1e764" containerName="kube-state-metrics" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544263 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0377d1a-719f-4733-bbd6-9cd135c1e764" containerName="kube-state-metrics" Jan 06 15:02:52 crc kubenswrapper[4744]: E0106 15:02:52.544279 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1890b419-b512-4925-993a-441edb67ba88" containerName="registry-server" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544287 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1890b419-b512-4925-993a-441edb67ba88" containerName="registry-server" Jan 06 15:02:52 crc kubenswrapper[4744]: E0106 15:02:52.544309 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1890b419-b512-4925-993a-441edb67ba88" containerName="extract-content" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544320 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1890b419-b512-4925-993a-441edb67ba88" containerName="extract-content" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544578 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1890b419-b512-4925-993a-441edb67ba88" containerName="registry-server" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544605 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe94fea2-a713-4982-8d58-9bd7176d99ed" containerName="mysqld-exporter" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544633 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0377d1a-719f-4733-bbd6-9cd135c1e764" containerName="kube-state-metrics" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.544648 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aea2ebb-3358-4d04-9f92-c4f2e43fb746" containerName="nova-cell1-novncproxy-novncproxy" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.545628 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.554053 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.554381 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.555645 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.567900 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.584473 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.594373 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.596290 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.605922 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.606136 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.607581 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.609514 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.614570 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.623361 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.623579 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.673791 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.673847 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d3de113-d71d-4f12-bebc-d3466483e800-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.673907 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d3de113-d71d-4f12-bebc-d3466483e800-config-data\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.673932 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.674004 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv7xw\" (UniqueName: \"kubernetes.io/projected/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-kube-api-access-vv7xw\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.674221 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwbrn\" (UniqueName: \"kubernetes.io/projected/6d3de113-d71d-4f12-bebc-d3466483e800-kube-api-access-bwbrn\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.674318 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d3de113-d71d-4f12-bebc-d3466483e800-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.674470 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.674520 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th6k4\" (UniqueName: \"kubernetes.io/projected/1999f288-0cad-400d-9c96-f073c424881b-kube-api-access-th6k4\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.674595 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.674647 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.674689 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.674731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.677670 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.773724 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777212 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777415 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777444 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777507 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777527 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d3de113-d71d-4f12-bebc-d3466483e800-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777569 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d3de113-d71d-4f12-bebc-d3466483e800-config-data\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777594 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv7xw\" (UniqueName: \"kubernetes.io/projected/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-kube-api-access-vv7xw\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777632 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwbrn\" (UniqueName: \"kubernetes.io/projected/6d3de113-d71d-4f12-bebc-d3466483e800-kube-api-access-bwbrn\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777663 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d3de113-d71d-4f12-bebc-d3466483e800-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777724 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.777760 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th6k4\" (UniqueName: \"kubernetes.io/projected/1999f288-0cad-400d-9c96-f073c424881b-kube-api-access-th6k4\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.796181 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d3de113-d71d-4f12-bebc-d3466483e800-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.803505 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.804223 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d3de113-d71d-4f12-bebc-d3466483e800-config-data\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.810216 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th6k4\" (UniqueName: \"kubernetes.io/projected/1999f288-0cad-400d-9c96-f073c424881b-kube-api-access-th6k4\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.810553 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.816725 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.817448 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.827198 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.834350 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d3de113-d71d-4f12-bebc-d3466483e800-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.843838 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.871444 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1999f288-0cad-400d-9c96-f073c424881b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1999f288-0cad-400d-9c96-f073c424881b\") " pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.872018 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwbrn\" (UniqueName: \"kubernetes.io/projected/6d3de113-d71d-4f12-bebc-d3466483e800-kube-api-access-bwbrn\") pod \"mysqld-exporter-0\" (UID: \"6d3de113-d71d-4f12-bebc-d3466483e800\") " pod="openstack/mysqld-exporter-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.882789 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv7xw\" (UniqueName: \"kubernetes.io/projected/a9ce0f2f-e4c5-4cc5-8196-0340d738f830-kube-api-access-vv7xw\") pod \"kube-state-metrics-0\" (UID: \"a9ce0f2f-e4c5-4cc5-8196-0340d738f830\") " pod="openstack/kube-state-metrics-0" Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.942526 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" event={"ID":"7405beaf-33c5-493b-bba7-83419a434632","Type":"ContainerStarted","Data":"a3d9e853b4458da1315369134ae769779b23055e524692501caed53c886918fa"} Jan 06 15:02:52 crc kubenswrapper[4744]: I0106 15:02:52.976845 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:53 crc kubenswrapper[4744]: I0106 15:02:53.008191 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 06 15:02:53 crc kubenswrapper[4744]: I0106 15:02:53.169945 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Jan 06 15:02:53 crc kubenswrapper[4744]: I0106 15:02:53.726496 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1890b419-b512-4925-993a-441edb67ba88" path="/var/lib/kubelet/pods/1890b419-b512-4925-993a-441edb67ba88/volumes" Jan 06 15:02:53 crc kubenswrapper[4744]: I0106 15:02:53.727848 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aea2ebb-3358-4d04-9f92-c4f2e43fb746" path="/var/lib/kubelet/pods/8aea2ebb-3358-4d04-9f92-c4f2e43fb746/volumes" Jan 06 15:02:53 crc kubenswrapper[4744]: I0106 15:02:53.728375 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0377d1a-719f-4733-bbd6-9cd135c1e764" path="/var/lib/kubelet/pods/d0377d1a-719f-4733-bbd6-9cd135c1e764/volumes" Jan 06 15:02:53 crc kubenswrapper[4744]: I0106 15:02:53.729383 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe94fea2-a713-4982-8d58-9bd7176d99ed" path="/var/lib/kubelet/pods/fe94fea2-a713-4982-8d58-9bd7176d99ed/volumes" Jan 06 15:02:53 crc kubenswrapper[4744]: I0106 15:02:53.786622 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 06 15:02:53 crc kubenswrapper[4744]: W0106 15:02:53.815838 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9ce0f2f_e4c5_4cc5_8196_0340d738f830.slice/crio-13c99d2de40afed6558b788a369253b111eb2e4c658318c7dc90064793755bc4 WatchSource:0}: Error finding container 13c99d2de40afed6558b788a369253b111eb2e4c658318c7dc90064793755bc4: Status 404 returned error can't find the container with id 13c99d2de40afed6558b788a369253b111eb2e4c658318c7dc90064793755bc4 Jan 06 15:02:53 crc kubenswrapper[4744]: I0106 15:02:53.816716 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 06 15:02:53 crc kubenswrapper[4744]: I0106 15:02:53.984834 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Jan 06 15:02:54 crc kubenswrapper[4744]: I0106 15:02:54.041971 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1999f288-0cad-400d-9c96-f073c424881b","Type":"ContainerStarted","Data":"b1c3b8928f7a10697e730f30fc247890d4afb871d5caddb3998a069b2fe6f028"} Jan 06 15:02:54 crc kubenswrapper[4744]: I0106 15:02:54.043008 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"6d3de113-d71d-4f12-bebc-d3466483e800","Type":"ContainerStarted","Data":"75b996f091dd7fc27711852320a69e95362771835efa2f34d9949973867e4259"} Jan 06 15:02:54 crc kubenswrapper[4744]: I0106 15:02:54.043907 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a9ce0f2f-e4c5-4cc5-8196-0340d738f830","Type":"ContainerStarted","Data":"13c99d2de40afed6558b788a369253b111eb2e4c658318c7dc90064793755bc4"} Jan 06 15:02:54 crc kubenswrapper[4744]: I0106 15:02:54.396537 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:02:54 crc kubenswrapper[4744]: I0106 15:02:54.396893 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-log" containerID="cri-o://cf11091fad049ead87f552949b247d2bb1d4a6b4902f4a64b34c79b25ea19bca" gracePeriod=30 Jan 06 15:02:54 crc kubenswrapper[4744]: I0106 15:02:54.397025 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-api" containerID="cri-o://98872de1df1609af11f2483bc9ef2043eb79b0e9b7b4037a7df854e2a9428182" gracePeriod=30 Jan 06 15:02:55 crc kubenswrapper[4744]: I0106 15:02:55.068897 4744 generic.go:334] "Generic (PLEG): container finished" podID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerID="cf11091fad049ead87f552949b247d2bb1d4a6b4902f4a64b34c79b25ea19bca" exitCode=143 Jan 06 15:02:55 crc kubenswrapper[4744]: I0106 15:02:55.068972 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"687c5ae1-31e8-4b4d-9f69-25f3095b10b1","Type":"ContainerDied","Data":"cf11091fad049ead87f552949b247d2bb1d4a6b4902f4a64b34c79b25ea19bca"} Jan 06 15:02:55 crc kubenswrapper[4744]: I0106 15:02:55.071863 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1999f288-0cad-400d-9c96-f073c424881b","Type":"ContainerStarted","Data":"af35faa3fadf973f66ed7befb6984d7412e02906a13ec8549ca4c0d0dded2ebc"} Jan 06 15:02:55 crc kubenswrapper[4744]: I0106 15:02:55.074624 4744 generic.go:334] "Generic (PLEG): container finished" podID="7405beaf-33c5-493b-bba7-83419a434632" containerID="aa6420ec614c010d558abaa6092c6de2db2940f60b5c9afc7f6a05f7a81c42da" exitCode=0 Jan 06 15:02:55 crc kubenswrapper[4744]: I0106 15:02:55.074662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" event={"ID":"7405beaf-33c5-493b-bba7-83419a434632","Type":"ContainerDied","Data":"aa6420ec614c010d558abaa6092c6de2db2940f60b5c9afc7f6a05f7a81c42da"} Jan 06 15:02:56 crc kubenswrapper[4744]: I0106 15:02:56.105538 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=4.105513191 podStartE2EDuration="4.105513191s" podCreationTimestamp="2026-01-06 15:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:02:56.10245776 +0000 UTC m=+1572.729924078" watchObservedRunningTime="2026-01-06 15:02:56.105513191 +0000 UTC m=+1572.732979529" Jan 06 15:02:57 crc kubenswrapper[4744]: I0106 15:02:57.977602 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.124491 4744 generic.go:334] "Generic (PLEG): container finished" podID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerID="98872de1df1609af11f2483bc9ef2043eb79b0e9b7b4037a7df854e2a9428182" exitCode=0 Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.124588 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"687c5ae1-31e8-4b4d-9f69-25f3095b10b1","Type":"ContainerDied","Data":"98872de1df1609af11f2483bc9ef2043eb79b0e9b7b4037a7df854e2a9428182"} Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.615106 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.689069 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-config-data\") pod \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.689243 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-logs\") pod \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.689280 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-combined-ca-bundle\") pod \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.689398 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7vmk\" (UniqueName: \"kubernetes.io/projected/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-kube-api-access-s7vmk\") pod \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\" (UID: \"687c5ae1-31e8-4b4d-9f69-25f3095b10b1\") " Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.690781 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-logs" (OuterVolumeSpecName: "logs") pod "687c5ae1-31e8-4b4d-9f69-25f3095b10b1" (UID: "687c5ae1-31e8-4b4d-9f69-25f3095b10b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.701284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-kube-api-access-s7vmk" (OuterVolumeSpecName: "kube-api-access-s7vmk") pod "687c5ae1-31e8-4b4d-9f69-25f3095b10b1" (UID: "687c5ae1-31e8-4b4d-9f69-25f3095b10b1"). InnerVolumeSpecName "kube-api-access-s7vmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.734934 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-config-data" (OuterVolumeSpecName: "config-data") pod "687c5ae1-31e8-4b4d-9f69-25f3095b10b1" (UID: "687c5ae1-31e8-4b4d-9f69-25f3095b10b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.744459 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "687c5ae1-31e8-4b4d-9f69-25f3095b10b1" (UID: "687c5ae1-31e8-4b4d-9f69-25f3095b10b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.794502 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.798580 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.798727 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:02:59 crc kubenswrapper[4744]: I0106 15:02:59.798785 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7vmk\" (UniqueName: \"kubernetes.io/projected/687c5ae1-31e8-4b4d-9f69-25f3095b10b1-kube-api-access-s7vmk\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.145086 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"687c5ae1-31e8-4b4d-9f69-25f3095b10b1","Type":"ContainerDied","Data":"37a41e0aef39f1305cb3fb39bc82468053bb560d84814f15c2ceae58e34c7e5c"} Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.145141 4744 scope.go:117] "RemoveContainer" containerID="98872de1df1609af11f2483bc9ef2043eb79b0e9b7b4037a7df854e2a9428182" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.145145 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.175623 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.193360 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.206861 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:00 crc kubenswrapper[4744]: E0106 15:03:00.207386 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-api" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.207398 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-api" Jan 06 15:03:00 crc kubenswrapper[4744]: E0106 15:03:00.207414 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-log" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.207419 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-log" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.207693 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-log" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.207714 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" containerName="nova-api-api" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.209132 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.213979 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.214492 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.218016 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.218297 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.309763 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.310009 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-logs\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.310104 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-public-tls-certs\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.310283 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-config-data\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.310405 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.310503 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9bs2\" (UniqueName: \"kubernetes.io/projected/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-kube-api-access-w9bs2\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.412914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.413145 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-logs\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.413263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-public-tls-certs\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.413429 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-config-data\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.413618 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.413717 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9bs2\" (UniqueName: \"kubernetes.io/projected/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-kube-api-access-w9bs2\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.413656 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-logs\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.417413 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-public-tls-certs\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.417554 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-config-data\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.419594 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.422517 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.440574 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9bs2\" (UniqueName: \"kubernetes.io/projected/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-kube-api-access-w9bs2\") pod \"nova-api-0\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " pod="openstack/nova-api-0" Jan 06 15:03:00 crc kubenswrapper[4744]: I0106 15:03:00.529578 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:03:01 crc kubenswrapper[4744]: I0106 15:03:01.732637 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="687c5ae1-31e8-4b4d-9f69-25f3095b10b1" path="/var/lib/kubelet/pods/687c5ae1-31e8-4b4d-9f69-25f3095b10b1/volumes" Jan 06 15:03:01 crc kubenswrapper[4744]: I0106 15:03:01.816757 4744 scope.go:117] "RemoveContainer" containerID="cf11091fad049ead87f552949b247d2bb1d4a6b4902f4a64b34c79b25ea19bca" Jan 06 15:03:02 crc kubenswrapper[4744]: I0106 15:03:02.923803 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:02 crc kubenswrapper[4744]: I0106 15:03:02.977629 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:03:03 crc kubenswrapper[4744]: I0106 15:03:03.758564 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:03:03 crc kubenswrapper[4744]: I0106 15:03:03.783080 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.263519 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e74e552-fedc-4e70-ab6c-2f1bb36e2011","Type":"ContainerStarted","Data":"25d46d3902912827169841c9f793f36ab947f89e358c150d2b63e9b4e328d82b"} Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.297535 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-8gfkz"] Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.299783 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.308574 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerStarted","Data":"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0"} Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.311562 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.311885 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.321486 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8gfkz"] Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.391985 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" event={"ID":"7405beaf-33c5-493b-bba7-83419a434632","Type":"ContainerStarted","Data":"e3637667911620da1180a235994b6954085165329abea4bb7a1f23fb84667847"} Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.392069 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.408348 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"6d3de113-d71d-4f12-bebc-d3466483e800","Type":"ContainerStarted","Data":"696e742a19db4df731d35da563b3b54742ee342db68b520eb4924de7b53cc404"} Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.436563 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" podStartSLOduration=14.436547353 podStartE2EDuration="14.436547353s" podCreationTimestamp="2026-01-06 15:02:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:03:04.425346426 +0000 UTC m=+1581.052812754" watchObservedRunningTime="2026-01-06 15:03:04.436547353 +0000 UTC m=+1581.064013661" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.459123 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=4.470666394 podStartE2EDuration="12.45910492s" podCreationTimestamp="2026-01-06 15:02:52 +0000 UTC" firstStartedPulling="2026-01-06 15:02:53.981840796 +0000 UTC m=+1570.609307114" lastFinishedPulling="2026-01-06 15:03:01.970279312 +0000 UTC m=+1578.597745640" observedRunningTime="2026-01-06 15:03:04.440802335 +0000 UTC m=+1581.068268663" watchObservedRunningTime="2026-01-06 15:03:04.45910492 +0000 UTC m=+1581.086571238" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.522148 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.522269 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-scripts\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.522340 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6h9w\" (UniqueName: \"kubernetes.io/projected/75eb54e3-63f3-4766-b95b-dbb1b307950c-kube-api-access-m6h9w\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.522449 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-config-data\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.624394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-scripts\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.624523 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6h9w\" (UniqueName: \"kubernetes.io/projected/75eb54e3-63f3-4766-b95b-dbb1b307950c-kube-api-access-m6h9w\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.624658 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-config-data\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.624804 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.636567 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-config-data\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.641517 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-scripts\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.645188 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.673795 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6h9w\" (UniqueName: \"kubernetes.io/projected/75eb54e3-63f3-4766-b95b-dbb1b307950c-kube-api-access-m6h9w\") pod \"nova-cell1-cell-mapping-8gfkz\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.712299 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:03:04 crc kubenswrapper[4744]: E0106 15:03:04.712592 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:03:04 crc kubenswrapper[4744]: I0106 15:03:04.953878 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:05 crc kubenswrapper[4744]: I0106 15:03:05.486318 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8gfkz"] Jan 06 15:03:05 crc kubenswrapper[4744]: W0106 15:03:05.486591 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75eb54e3_63f3_4766_b95b_dbb1b307950c.slice/crio-bf5acb8309fdd529956a257ccc0c5ba82c35a1941cb030eb83e3aef599e158f6 WatchSource:0}: Error finding container bf5acb8309fdd529956a257ccc0c5ba82c35a1941cb030eb83e3aef599e158f6: Status 404 returned error can't find the container with id bf5acb8309fdd529956a257ccc0c5ba82c35a1941cb030eb83e3aef599e158f6 Jan 06 15:03:06 crc kubenswrapper[4744]: I0106 15:03:06.447118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"a9ce0f2f-e4c5-4cc5-8196-0340d738f830","Type":"ContainerStarted","Data":"15ffeb21a494f6b5c9a463c5f2bbbe505b7a337771cdb0b7d704d0c707a6a950"} Jan 06 15:03:06 crc kubenswrapper[4744]: I0106 15:03:06.448814 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8gfkz" event={"ID":"75eb54e3-63f3-4766-b95b-dbb1b307950c","Type":"ContainerStarted","Data":"bf5acb8309fdd529956a257ccc0c5ba82c35a1941cb030eb83e3aef599e158f6"} Jan 06 15:03:06 crc kubenswrapper[4744]: I0106 15:03:06.451467 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerStarted","Data":"ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595"} Jan 06 15:03:07 crc kubenswrapper[4744]: I0106 15:03:07.470323 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e74e552-fedc-4e70-ab6c-2f1bb36e2011","Type":"ContainerStarted","Data":"cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a"} Jan 06 15:03:07 crc kubenswrapper[4744]: I0106 15:03:07.471752 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e74e552-fedc-4e70-ab6c-2f1bb36e2011","Type":"ContainerStarted","Data":"0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef"} Jan 06 15:03:07 crc kubenswrapper[4744]: I0106 15:03:07.475352 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerStarted","Data":"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5"} Jan 06 15:03:07 crc kubenswrapper[4744]: I0106 15:03:07.481901 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8gfkz" event={"ID":"75eb54e3-63f3-4766-b95b-dbb1b307950c","Type":"ContainerStarted","Data":"d268bae99d5a17ddfa477ccd0a637090326bbeff11dba3c3456b8a3579abdc86"} Jan 06 15:03:07 crc kubenswrapper[4744]: I0106 15:03:07.481951 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 06 15:03:07 crc kubenswrapper[4744]: I0106 15:03:07.512742 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=7.512724193 podStartE2EDuration="7.512724193s" podCreationTimestamp="2026-01-06 15:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:03:07.510247017 +0000 UTC m=+1584.137713355" watchObservedRunningTime="2026-01-06 15:03:07.512724193 +0000 UTC m=+1584.140190511" Jan 06 15:03:07 crc kubenswrapper[4744]: I0106 15:03:07.555733 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-8gfkz" podStartSLOduration=3.55571149 podStartE2EDuration="3.55571149s" podCreationTimestamp="2026-01-06 15:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:03:07.539491171 +0000 UTC m=+1584.166957509" watchObservedRunningTime="2026-01-06 15:03:07.55571149 +0000 UTC m=+1584.183177808" Jan 06 15:03:07 crc kubenswrapper[4744]: I0106 15:03:07.580317 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=7.059152538 podStartE2EDuration="15.580296201s" podCreationTimestamp="2026-01-06 15:02:52 +0000 UTC" firstStartedPulling="2026-01-06 15:02:53.823401923 +0000 UTC m=+1570.450868241" lastFinishedPulling="2026-01-06 15:03:02.344545566 +0000 UTC m=+1578.972011904" observedRunningTime="2026-01-06 15:03:07.561937285 +0000 UTC m=+1584.189403603" watchObservedRunningTime="2026-01-06 15:03:07.580296201 +0000 UTC m=+1584.207762519" Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.518525 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerStarted","Data":"700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827"} Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.518878 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-api" containerID="cri-o://8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5" gracePeriod=30 Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.519195 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-notifier" containerID="cri-o://ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595" gracePeriod=30 Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.519243 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-listener" containerID="cri-o://700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827" gracePeriod=30 Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.519294 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-evaluator" containerID="cri-o://e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5" gracePeriod=30 Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.530346 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.530384 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.531973 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="ceilometer-central-agent" containerID="cri-o://0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a" gracePeriod=30 Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.532071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerStarted","Data":"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad"} Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.532115 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.532264 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="proxy-httpd" containerID="cri-o://0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad" gracePeriod=30 Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.532317 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="sg-core" containerID="cri-o://c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5" gracePeriod=30 Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.532355 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="ceilometer-notification-agent" containerID="cri-o://ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0" gracePeriod=30 Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.560908 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.203444096 podStartE2EDuration="37.560886031s" podCreationTimestamp="2026-01-06 15:02:33 +0000 UTC" firstStartedPulling="2026-01-06 15:02:34.735188425 +0000 UTC m=+1551.362654743" lastFinishedPulling="2026-01-06 15:03:09.09263036 +0000 UTC m=+1585.720096678" observedRunningTime="2026-01-06 15:03:10.543726137 +0000 UTC m=+1587.171192475" watchObservedRunningTime="2026-01-06 15:03:10.560886031 +0000 UTC m=+1587.188352349" Jan 06 15:03:10 crc kubenswrapper[4744]: I0106 15:03:10.582931 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.147292789 podStartE2EDuration="29.582908304s" podCreationTimestamp="2026-01-06 15:02:41 +0000 UTC" firstStartedPulling="2026-01-06 15:02:42.658265788 +0000 UTC m=+1559.285732106" lastFinishedPulling="2026-01-06 15:03:09.093881303 +0000 UTC m=+1585.721347621" observedRunningTime="2026-01-06 15:03:10.56764614 +0000 UTC m=+1587.195112478" watchObservedRunningTime="2026-01-06 15:03:10.582908304 +0000 UTC m=+1587.210374622" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.254328 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.294255 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.376571 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-4wt4q"] Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.377022 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" podUID="aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" containerName="dnsmasq-dns" containerID="cri-o://e4392110e5cf7cd9c8d6dfb40853a304682005ec5f8a3eb9985b3fe0784ec548" gracePeriod=10 Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.420937 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-sg-core-conf-yaml\") pod \"89850656-f2cd-413f-9a79-9b08f80c129f\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.421009 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-run-httpd\") pod \"89850656-f2cd-413f-9a79-9b08f80c129f\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.421325 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-log-httpd\") pod \"89850656-f2cd-413f-9a79-9b08f80c129f\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.421396 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "89850656-f2cd-413f-9a79-9b08f80c129f" (UID: "89850656-f2cd-413f-9a79-9b08f80c129f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.421456 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-combined-ca-bundle\") pod \"89850656-f2cd-413f-9a79-9b08f80c129f\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.421500 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-scripts\") pod \"89850656-f2cd-413f-9a79-9b08f80c129f\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.421567 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-config-data\") pod \"89850656-f2cd-413f-9a79-9b08f80c129f\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.421593 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxpzp\" (UniqueName: \"kubernetes.io/projected/89850656-f2cd-413f-9a79-9b08f80c129f-kube-api-access-bxpzp\") pod \"89850656-f2cd-413f-9a79-9b08f80c129f\" (UID: \"89850656-f2cd-413f-9a79-9b08f80c129f\") " Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.421990 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "89850656-f2cd-413f-9a79-9b08f80c129f" (UID: "89850656-f2cd-413f-9a79-9b08f80c129f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.422689 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.422709 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/89850656-f2cd-413f-9a79-9b08f80c129f-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.427927 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-scripts" (OuterVolumeSpecName: "scripts") pod "89850656-f2cd-413f-9a79-9b08f80c129f" (UID: "89850656-f2cd-413f-9a79-9b08f80c129f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.428933 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89850656-f2cd-413f-9a79-9b08f80c129f-kube-api-access-bxpzp" (OuterVolumeSpecName: "kube-api-access-bxpzp") pod "89850656-f2cd-413f-9a79-9b08f80c129f" (UID: "89850656-f2cd-413f-9a79-9b08f80c129f"). InnerVolumeSpecName "kube-api-access-bxpzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.495042 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "89850656-f2cd-413f-9a79-9b08f80c129f" (UID: "89850656-f2cd-413f-9a79-9b08f80c129f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.524868 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.524917 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.524927 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxpzp\" (UniqueName: \"kubernetes.io/projected/89850656-f2cd-413f-9a79-9b08f80c129f-kube-api-access-bxpzp\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.546506 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.5:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.546776 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.5:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.551275 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-config-data" (OuterVolumeSpecName: "config-data") pod "89850656-f2cd-413f-9a79-9b08f80c129f" (UID: "89850656-f2cd-413f-9a79-9b08f80c129f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.566044 4744 generic.go:334] "Generic (PLEG): container finished" podID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerID="e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5" exitCode=0 Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.566083 4744 generic.go:334] "Generic (PLEG): container finished" podID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerID="8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5" exitCode=0 Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.566134 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerDied","Data":"e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5"} Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.566182 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerDied","Data":"8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5"} Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.585803 4744 generic.go:334] "Generic (PLEG): container finished" podID="aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" containerID="e4392110e5cf7cd9c8d6dfb40853a304682005ec5f8a3eb9985b3fe0784ec548" exitCode=0 Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.585873 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" event={"ID":"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2","Type":"ContainerDied","Data":"e4392110e5cf7cd9c8d6dfb40853a304682005ec5f8a3eb9985b3fe0784ec548"} Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.591468 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89850656-f2cd-413f-9a79-9b08f80c129f" (UID: "89850656-f2cd-413f-9a79-9b08f80c129f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599045 4744 generic.go:334] "Generic (PLEG): container finished" podID="89850656-f2cd-413f-9a79-9b08f80c129f" containerID="0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad" exitCode=0 Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599077 4744 generic.go:334] "Generic (PLEG): container finished" podID="89850656-f2cd-413f-9a79-9b08f80c129f" containerID="c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5" exitCode=2 Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599084 4744 generic.go:334] "Generic (PLEG): container finished" podID="89850656-f2cd-413f-9a79-9b08f80c129f" containerID="ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0" exitCode=0 Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599092 4744 generic.go:334] "Generic (PLEG): container finished" podID="89850656-f2cd-413f-9a79-9b08f80c129f" containerID="0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a" exitCode=0 Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599114 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerDied","Data":"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad"} Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599153 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerDied","Data":"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5"} Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599177 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerDied","Data":"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0"} Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599188 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerDied","Data":"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a"} Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599197 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"89850656-f2cd-413f-9a79-9b08f80c129f","Type":"ContainerDied","Data":"5a9cbcc5d5ad34b1816c9fe3d1941cfd57b5074ef5d36f477342cc74028c2e79"} Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599214 4744 scope.go:117] "RemoveContainer" containerID="0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.599384 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.627014 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.627040 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89850656-f2cd-413f-9a79-9b08f80c129f-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.671049 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.681245 4744 scope.go:117] "RemoveContainer" containerID="c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.700313 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.846768 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" path="/var/lib/kubelet/pods/89850656-f2cd-413f-9a79-9b08f80c129f/volumes" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.847893 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:03:11 crc kubenswrapper[4744]: E0106 15:03:11.848768 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="ceilometer-central-agent" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.848784 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="ceilometer-central-agent" Jan 06 15:03:11 crc kubenswrapper[4744]: E0106 15:03:11.848807 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="sg-core" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.848814 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="sg-core" Jan 06 15:03:11 crc kubenswrapper[4744]: E0106 15:03:11.848841 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="ceilometer-notification-agent" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.848848 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="ceilometer-notification-agent" Jan 06 15:03:11 crc kubenswrapper[4744]: E0106 15:03:11.848968 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="proxy-httpd" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.849005 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="proxy-httpd" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.850170 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="sg-core" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.850199 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="proxy-httpd" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.850264 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="ceilometer-central-agent" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.850295 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="89850656-f2cd-413f-9a79-9b08f80c129f" containerName="ceilometer-notification-agent" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.854359 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.854487 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.859349 4744 scope.go:117] "RemoveContainer" containerID="ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.905317 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.906315 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 15:03:11 crc kubenswrapper[4744]: I0106 15:03:11.906489 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.000864 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-config-data\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.000989 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rx8w\" (UniqueName: \"kubernetes.io/projected/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-kube-api-access-2rx8w\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.001034 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.001096 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.001169 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-log-httpd\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.001199 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-scripts\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.001249 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.001279 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-run-httpd\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.001420 4744 scope.go:117] "RemoveContainer" containerID="0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.065343 4744 scope.go:117] "RemoveContainer" containerID="0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad" Jan 06 15:03:12 crc kubenswrapper[4744]: E0106 15:03:12.068087 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad\": container with ID starting with 0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad not found: ID does not exist" containerID="0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.068125 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad"} err="failed to get container status \"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad\": rpc error: code = NotFound desc = could not find container \"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad\": container with ID starting with 0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.068147 4744 scope.go:117] "RemoveContainer" containerID="c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5" Jan 06 15:03:12 crc kubenswrapper[4744]: E0106 15:03:12.069466 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5\": container with ID starting with c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5 not found: ID does not exist" containerID="c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.069509 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5"} err="failed to get container status \"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5\": rpc error: code = NotFound desc = could not find container \"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5\": container with ID starting with c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5 not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.069535 4744 scope.go:117] "RemoveContainer" containerID="ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0" Jan 06 15:03:12 crc kubenswrapper[4744]: E0106 15:03:12.070070 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0\": container with ID starting with ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0 not found: ID does not exist" containerID="ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070095 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0"} err="failed to get container status \"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0\": rpc error: code = NotFound desc = could not find container \"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0\": container with ID starting with ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0 not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070109 4744 scope.go:117] "RemoveContainer" containerID="0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a" Jan 06 15:03:12 crc kubenswrapper[4744]: E0106 15:03:12.070291 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a\": container with ID starting with 0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a not found: ID does not exist" containerID="0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070308 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a"} err="failed to get container status \"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a\": rpc error: code = NotFound desc = could not find container \"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a\": container with ID starting with 0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070320 4744 scope.go:117] "RemoveContainer" containerID="0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070466 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad"} err="failed to get container status \"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad\": rpc error: code = NotFound desc = could not find container \"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad\": container with ID starting with 0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070478 4744 scope.go:117] "RemoveContainer" containerID="c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070628 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5"} err="failed to get container status \"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5\": rpc error: code = NotFound desc = could not find container \"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5\": container with ID starting with c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5 not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070642 4744 scope.go:117] "RemoveContainer" containerID="ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070788 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0"} err="failed to get container status \"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0\": rpc error: code = NotFound desc = could not find container \"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0\": container with ID starting with ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0 not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070803 4744 scope.go:117] "RemoveContainer" containerID="0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070953 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a"} err="failed to get container status \"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a\": rpc error: code = NotFound desc = could not find container \"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a\": container with ID starting with 0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.070966 4744 scope.go:117] "RemoveContainer" containerID="0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.071310 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad"} err="failed to get container status \"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad\": rpc error: code = NotFound desc = could not find container \"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad\": container with ID starting with 0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.071324 4744 scope.go:117] "RemoveContainer" containerID="c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.071505 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5"} err="failed to get container status \"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5\": rpc error: code = NotFound desc = could not find container \"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5\": container with ID starting with c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5 not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.071519 4744 scope.go:117] "RemoveContainer" containerID="ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.071917 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0"} err="failed to get container status \"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0\": rpc error: code = NotFound desc = could not find container \"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0\": container with ID starting with ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0 not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.071936 4744 scope.go:117] "RemoveContainer" containerID="0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.086757 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a"} err="failed to get container status \"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a\": rpc error: code = NotFound desc = could not find container \"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a\": container with ID starting with 0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.086807 4744 scope.go:117] "RemoveContainer" containerID="0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.087917 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad"} err="failed to get container status \"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad\": rpc error: code = NotFound desc = could not find container \"0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad\": container with ID starting with 0769626beaa9aff91ad3ebba6ad9248e70ec61b59f3322bc6ce6276c9e4af4ad not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.087964 4744 scope.go:117] "RemoveContainer" containerID="c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.088736 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5"} err="failed to get container status \"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5\": rpc error: code = NotFound desc = could not find container \"c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5\": container with ID starting with c2886193f05a742b8d3ac02956ff6ea05067c4f175467b8b7df751ed1281fdc5 not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.088789 4744 scope.go:117] "RemoveContainer" containerID="ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.090579 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0"} err="failed to get container status \"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0\": rpc error: code = NotFound desc = could not find container \"ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0\": container with ID starting with ffc107d685b2ffaa50a931b6f486193c0304be5458f5918024f96ae3e1844cf0 not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.090602 4744 scope.go:117] "RemoveContainer" containerID="0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.091612 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a"} err="failed to get container status \"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a\": rpc error: code = NotFound desc = could not find container \"0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a\": container with ID starting with 0c1ec4b9d6f2b1c6bbaf830483cb63502193d80e7ab0a4b4dfca9d9e4b59003a not found: ID does not exist" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.104054 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rx8w\" (UniqueName: \"kubernetes.io/projected/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-kube-api-access-2rx8w\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.104172 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.104237 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.104270 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-log-httpd\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.104310 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-scripts\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.104385 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.104423 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-run-httpd\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.104731 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-config-data\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.104852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-log-httpd\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.107882 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-run-httpd\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.113089 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.120791 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-config-data\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.129017 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.131720 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-scripts\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.140233 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rx8w\" (UniqueName: \"kubernetes.io/projected/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-kube-api-access-2rx8w\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.140267 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.250141 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.320553 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.425251 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-config\") pod \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.425637 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-svc\") pod \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.425702 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-nb\") pod \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.425761 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j4v7\" (UniqueName: \"kubernetes.io/projected/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-kube-api-access-2j4v7\") pod \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.425784 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-swift-storage-0\") pod \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.425929 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-sb\") pod \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\" (UID: \"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2\") " Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.453231 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-kube-api-access-2j4v7" (OuterVolumeSpecName: "kube-api-access-2j4v7") pod "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" (UID: "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2"). InnerVolumeSpecName "kube-api-access-2j4v7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.509195 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-config" (OuterVolumeSpecName: "config") pod "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" (UID: "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.517838 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" (UID: "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.524503 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" (UID: "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.529409 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.529428 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j4v7\" (UniqueName: \"kubernetes.io/projected/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-kube-api-access-2j4v7\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.529439 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.529449 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.556875 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" (UID: "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.561607 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" (UID: "aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.621018 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.621016 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7877d89589-4wt4q" event={"ID":"aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2","Type":"ContainerDied","Data":"335a533282b99963b0fcd9f0ab9a9655a39699aee981b6537abbc48307a15a98"} Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.621191 4744 scope.go:117] "RemoveContainer" containerID="e4392110e5cf7cd9c8d6dfb40853a304682005ec5f8a3eb9985b3fe0784ec548" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.631077 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.631482 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.631843 4744 generic.go:334] "Generic (PLEG): container finished" podID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerID="ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595" exitCode=0 Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.631880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerDied","Data":"ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595"} Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.659364 4744 scope.go:117] "RemoveContainer" containerID="133dcfe7ccf5884e7b70cd789c452081c59268bb258bafdf957ceb6c07b9b6b6" Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.691914 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-4wt4q"] Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.702434 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7877d89589-4wt4q"] Jan 06 15:03:12 crc kubenswrapper[4744]: I0106 15:03:12.846687 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:03:13 crc kubenswrapper[4744]: I0106 15:03:13.016453 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 06 15:03:13 crc kubenswrapper[4744]: I0106 15:03:13.679207 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerStarted","Data":"3b51beadcd1518d0e7fc3ff6b4e556268cfda31dde8079af44abfe50eba438af"} Jan 06 15:03:13 crc kubenswrapper[4744]: I0106 15:03:13.732751 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" path="/var/lib/kubelet/pods/aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2/volumes" Jan 06 15:03:14 crc kubenswrapper[4744]: I0106 15:03:14.698414 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerStarted","Data":"f9fda546f2326b3980c08b17525d568a5f2c7fb9aa7958d4773c7010644a0d7e"} Jan 06 15:03:14 crc kubenswrapper[4744]: I0106 15:03:14.702429 4744 generic.go:334] "Generic (PLEG): container finished" podID="75eb54e3-63f3-4766-b95b-dbb1b307950c" containerID="d268bae99d5a17ddfa477ccd0a637090326bbeff11dba3c3456b8a3579abdc86" exitCode=0 Jan 06 15:03:14 crc kubenswrapper[4744]: I0106 15:03:14.702473 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8gfkz" event={"ID":"75eb54e3-63f3-4766-b95b-dbb1b307950c","Type":"ContainerDied","Data":"d268bae99d5a17ddfa477ccd0a637090326bbeff11dba3c3456b8a3579abdc86"} Jan 06 15:03:15 crc kubenswrapper[4744]: I0106 15:03:15.726492 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerStarted","Data":"7555282a4e1ae18eaf7b74f6e033056f94245ebf901084f171fedc45d006238a"} Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.169832 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.264927 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-combined-ca-bundle\") pod \"75eb54e3-63f3-4766-b95b-dbb1b307950c\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.265252 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-scripts\") pod \"75eb54e3-63f3-4766-b95b-dbb1b307950c\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.265285 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-config-data\") pod \"75eb54e3-63f3-4766-b95b-dbb1b307950c\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.265390 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m6h9w\" (UniqueName: \"kubernetes.io/projected/75eb54e3-63f3-4766-b95b-dbb1b307950c-kube-api-access-m6h9w\") pod \"75eb54e3-63f3-4766-b95b-dbb1b307950c\" (UID: \"75eb54e3-63f3-4766-b95b-dbb1b307950c\") " Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.271273 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-scripts" (OuterVolumeSpecName: "scripts") pod "75eb54e3-63f3-4766-b95b-dbb1b307950c" (UID: "75eb54e3-63f3-4766-b95b-dbb1b307950c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.271324 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75eb54e3-63f3-4766-b95b-dbb1b307950c-kube-api-access-m6h9w" (OuterVolumeSpecName: "kube-api-access-m6h9w") pod "75eb54e3-63f3-4766-b95b-dbb1b307950c" (UID: "75eb54e3-63f3-4766-b95b-dbb1b307950c"). InnerVolumeSpecName "kube-api-access-m6h9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.298294 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-config-data" (OuterVolumeSpecName: "config-data") pod "75eb54e3-63f3-4766-b95b-dbb1b307950c" (UID: "75eb54e3-63f3-4766-b95b-dbb1b307950c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.304679 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75eb54e3-63f3-4766-b95b-dbb1b307950c" (UID: "75eb54e3-63f3-4766-b95b-dbb1b307950c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.369067 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.369116 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.369131 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m6h9w\" (UniqueName: \"kubernetes.io/projected/75eb54e3-63f3-4766-b95b-dbb1b307950c-kube-api-access-m6h9w\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.369146 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75eb54e3-63f3-4766-b95b-dbb1b307950c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.735697 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8gfkz" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.737723 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8gfkz" event={"ID":"75eb54e3-63f3-4766-b95b-dbb1b307950c","Type":"ContainerDied","Data":"bf5acb8309fdd529956a257ccc0c5ba82c35a1941cb030eb83e3aef599e158f6"} Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.737809 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf5acb8309fdd529956a257ccc0c5ba82c35a1941cb030eb83e3aef599e158f6" Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.742850 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerStarted","Data":"ceea7c24813d0da0f03e90e5acfd4247ca59fed292baa7666a8555ee19da01cb"} Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.927289 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.927535 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c37294a5-5d6b-495a-a392-7744df1c0ec9" containerName="nova-scheduler-scheduler" containerID="cri-o://3d1a9b23a5439d8d400ab9bc66ce612d21b70c49b9b82e16f6d214f81fe399b8" gracePeriod=30 Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.988182 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.988462 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-log" containerID="cri-o://0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef" gracePeriod=30 Jan 06 15:03:16 crc kubenswrapper[4744]: I0106 15:03:16.988803 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-api" containerID="cri-o://cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a" gracePeriod=30 Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.002294 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.002525 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-log" containerID="cri-o://96d4d2f7c18c6829a3a079d4f8c558a32b22a7e809303f87b78955a94a9a568f" gracePeriod=30 Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.002576 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-metadata" containerID="cri-o://a88070ac4ba00f8e117f58818e579e3cc3c8cad215bcc3b69647a07f2884409f" gracePeriod=30 Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.711849 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:03:17 crc kubenswrapper[4744]: E0106 15:03:17.712353 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.761668 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerStarted","Data":"481493d2e869b5d33bc577565682801a87433163305e82bb60062f4559087d28"} Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.761800 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.765670 4744 generic.go:334] "Generic (PLEG): container finished" podID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerID="96d4d2f7c18c6829a3a079d4f8c558a32b22a7e809303f87b78955a94a9a568f" exitCode=143 Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.765752 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9255f27d-3d50-47f7-92b0-ee9331a565cd","Type":"ContainerDied","Data":"96d4d2f7c18c6829a3a079d4f8c558a32b22a7e809303f87b78955a94a9a568f"} Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.775563 4744 generic.go:334] "Generic (PLEG): container finished" podID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerID="0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef" exitCode=143 Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.775621 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e74e552-fedc-4e70-ab6c-2f1bb36e2011","Type":"ContainerDied","Data":"0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef"} Jan 06 15:03:17 crc kubenswrapper[4744]: I0106 15:03:17.792533 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.368083364 podStartE2EDuration="6.792509681s" podCreationTimestamp="2026-01-06 15:03:11 +0000 UTC" firstStartedPulling="2026-01-06 15:03:12.867441207 +0000 UTC m=+1589.494907525" lastFinishedPulling="2026-01-06 15:03:17.291867524 +0000 UTC m=+1593.919333842" observedRunningTime="2026-01-06 15:03:17.785509285 +0000 UTC m=+1594.412975603" watchObservedRunningTime="2026-01-06 15:03:17.792509681 +0000 UTC m=+1594.419975999" Jan 06 15:03:18 crc kubenswrapper[4744]: E0106 15:03:18.588091 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc37294a5_5d6b_495a_a392_7744df1c0ec9.slice/crio-conmon-3d1a9b23a5439d8d400ab9bc66ce612d21b70c49b9b82e16f6d214f81fe399b8.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:03:18 crc kubenswrapper[4744]: I0106 15:03:18.794695 4744 generic.go:334] "Generic (PLEG): container finished" podID="c37294a5-5d6b-495a-a392-7744df1c0ec9" containerID="3d1a9b23a5439d8d400ab9bc66ce612d21b70c49b9b82e16f6d214f81fe399b8" exitCode=0 Jan 06 15:03:18 crc kubenswrapper[4744]: I0106 15:03:18.794755 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c37294a5-5d6b-495a-a392-7744df1c0ec9","Type":"ContainerDied","Data":"3d1a9b23a5439d8d400ab9bc66ce612d21b70c49b9b82e16f6d214f81fe399b8"} Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.025206 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.144113 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-config-data\") pod \"c37294a5-5d6b-495a-a392-7744df1c0ec9\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.144260 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-combined-ca-bundle\") pod \"c37294a5-5d6b-495a-a392-7744df1c0ec9\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.144305 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcn2p\" (UniqueName: \"kubernetes.io/projected/c37294a5-5d6b-495a-a392-7744df1c0ec9-kube-api-access-xcn2p\") pod \"c37294a5-5d6b-495a-a392-7744df1c0ec9\" (UID: \"c37294a5-5d6b-495a-a392-7744df1c0ec9\") " Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.155236 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c37294a5-5d6b-495a-a392-7744df1c0ec9-kube-api-access-xcn2p" (OuterVolumeSpecName: "kube-api-access-xcn2p") pod "c37294a5-5d6b-495a-a392-7744df1c0ec9" (UID: "c37294a5-5d6b-495a-a392-7744df1c0ec9"). InnerVolumeSpecName "kube-api-access-xcn2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.176740 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-config-data" (OuterVolumeSpecName: "config-data") pod "c37294a5-5d6b-495a-a392-7744df1c0ec9" (UID: "c37294a5-5d6b-495a-a392-7744df1c0ec9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.181451 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c37294a5-5d6b-495a-a392-7744df1c0ec9" (UID: "c37294a5-5d6b-495a-a392-7744df1c0ec9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.246500 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.246538 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c37294a5-5d6b-495a-a392-7744df1c0ec9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.246550 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcn2p\" (UniqueName: \"kubernetes.io/projected/c37294a5-5d6b-495a-a392-7744df1c0ec9-kube-api-access-xcn2p\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.818924 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c37294a5-5d6b-495a-a392-7744df1c0ec9","Type":"ContainerDied","Data":"512fb1f7b14c7fa98c4928e2822ce2cfc62b0640e93d503d1355b51fe382fc95"} Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.818979 4744 scope.go:117] "RemoveContainer" containerID="3d1a9b23a5439d8d400ab9bc66ce612d21b70c49b9b82e16f6d214f81fe399b8" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.818984 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.860633 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.892320 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.918476 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:03:19 crc kubenswrapper[4744]: E0106 15:03:19.919236 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" containerName="init" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.919254 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" containerName="init" Jan 06 15:03:19 crc kubenswrapper[4744]: E0106 15:03:19.919267 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" containerName="dnsmasq-dns" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.919273 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" containerName="dnsmasq-dns" Jan 06 15:03:19 crc kubenswrapper[4744]: E0106 15:03:19.920408 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c37294a5-5d6b-495a-a392-7744df1c0ec9" containerName="nova-scheduler-scheduler" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.920423 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c37294a5-5d6b-495a-a392-7744df1c0ec9" containerName="nova-scheduler-scheduler" Jan 06 15:03:19 crc kubenswrapper[4744]: E0106 15:03:19.920446 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75eb54e3-63f3-4766-b95b-dbb1b307950c" containerName="nova-manage" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.920451 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="75eb54e3-63f3-4766-b95b-dbb1b307950c" containerName="nova-manage" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.920726 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="75eb54e3-63f3-4766-b95b-dbb1b307950c" containerName="nova-manage" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.920776 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c37294a5-5d6b-495a-a392-7744df1c0ec9" containerName="nova-scheduler-scheduler" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.920797 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa57fab9-b04f-4a72-ba2f-2ee0e76e56b2" containerName="dnsmasq-dns" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.922189 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.927646 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 06 15:03:19 crc kubenswrapper[4744]: I0106 15:03:19.932073 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.069696 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8221b070-0d54-439f-887a-9a401d634bea-config-data\") pod \"nova-scheduler-0\" (UID: \"8221b070-0d54-439f-887a-9a401d634bea\") " pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.069748 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8221b070-0d54-439f-887a-9a401d634bea-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8221b070-0d54-439f-887a-9a401d634bea\") " pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.069914 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvlpf\" (UniqueName: \"kubernetes.io/projected/8221b070-0d54-439f-887a-9a401d634bea-kube-api-access-lvlpf\") pod \"nova-scheduler-0\" (UID: \"8221b070-0d54-439f-887a-9a401d634bea\") " pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.172389 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvlpf\" (UniqueName: \"kubernetes.io/projected/8221b070-0d54-439f-887a-9a401d634bea-kube-api-access-lvlpf\") pod \"nova-scheduler-0\" (UID: \"8221b070-0d54-439f-887a-9a401d634bea\") " pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.172558 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8221b070-0d54-439f-887a-9a401d634bea-config-data\") pod \"nova-scheduler-0\" (UID: \"8221b070-0d54-439f-887a-9a401d634bea\") " pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.172606 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8221b070-0d54-439f-887a-9a401d634bea-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8221b070-0d54-439f-887a-9a401d634bea\") " pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.176622 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8221b070-0d54-439f-887a-9a401d634bea-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8221b070-0d54-439f-887a-9a401d634bea\") " pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.176831 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8221b070-0d54-439f-887a-9a401d634bea-config-data\") pod \"nova-scheduler-0\" (UID: \"8221b070-0d54-439f-887a-9a401d634bea\") " pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.201226 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvlpf\" (UniqueName: \"kubernetes.io/projected/8221b070-0d54-439f-887a-9a401d634bea-kube-api-access-lvlpf\") pod \"nova-scheduler-0\" (UID: \"8221b070-0d54-439f-887a-9a401d634bea\") " pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.244368 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.425323 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.251:8775/\": read tcp 10.217.0.2:47988->10.217.0.251:8775: read: connection reset by peer" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.425618 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.251:8775/\": read tcp 10.217.0.2:47990->10.217.0.251:8775: read: connection reset by peer" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.727316 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.787201 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-config-data\") pod \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.787257 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-combined-ca-bundle\") pod \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.787355 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9bs2\" (UniqueName: \"kubernetes.io/projected/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-kube-api-access-w9bs2\") pod \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.787423 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-public-tls-certs\") pod \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.787521 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-internal-tls-certs\") pod \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.787605 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-logs\") pod \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\" (UID: \"0e74e552-fedc-4e70-ab6c-2f1bb36e2011\") " Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.789284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-logs" (OuterVolumeSpecName: "logs") pod "0e74e552-fedc-4e70-ab6c-2f1bb36e2011" (UID: "0e74e552-fedc-4e70-ab6c-2f1bb36e2011"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.804889 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-kube-api-access-w9bs2" (OuterVolumeSpecName: "kube-api-access-w9bs2") pod "0e74e552-fedc-4e70-ab6c-2f1bb36e2011" (UID: "0e74e552-fedc-4e70-ab6c-2f1bb36e2011"). InnerVolumeSpecName "kube-api-access-w9bs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.838640 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-config-data" (OuterVolumeSpecName: "config-data") pod "0e74e552-fedc-4e70-ab6c-2f1bb36e2011" (UID: "0e74e552-fedc-4e70-ab6c-2f1bb36e2011"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.846331 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e74e552-fedc-4e70-ab6c-2f1bb36e2011" (UID: "0e74e552-fedc-4e70-ab6c-2f1bb36e2011"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:20 crc kubenswrapper[4744]: W0106 15:03:20.856879 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8221b070_0d54_439f_887a_9a401d634bea.slice/crio-1dc0dec304aaf7cd8c29856307353cf2d2778c227c9e444bc32046889b23f6ce WatchSource:0}: Error finding container 1dc0dec304aaf7cd8c29856307353cf2d2778c227c9e444bc32046889b23f6ce: Status 404 returned error can't find the container with id 1dc0dec304aaf7cd8c29856307353cf2d2778c227c9e444bc32046889b23f6ce Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.858119 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.858949 4744 generic.go:334] "Generic (PLEG): container finished" podID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerID="a88070ac4ba00f8e117f58818e579e3cc3c8cad215bcc3b69647a07f2884409f" exitCode=0 Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.859011 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9255f27d-3d50-47f7-92b0-ee9331a565cd","Type":"ContainerDied","Data":"a88070ac4ba00f8e117f58818e579e3cc3c8cad215bcc3b69647a07f2884409f"} Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.864978 4744 generic.go:334] "Generic (PLEG): container finished" podID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerID="cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a" exitCode=0 Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.865024 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e74e552-fedc-4e70-ab6c-2f1bb36e2011","Type":"ContainerDied","Data":"cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a"} Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.865061 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e74e552-fedc-4e70-ab6c-2f1bb36e2011","Type":"ContainerDied","Data":"25d46d3902912827169841c9f793f36ab947f89e358c150d2b63e9b4e328d82b"} Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.865082 4744 scope.go:117] "RemoveContainer" containerID="cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.865252 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.868693 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0e74e552-fedc-4e70-ab6c-2f1bb36e2011" (UID: "0e74e552-fedc-4e70-ab6c-2f1bb36e2011"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.892697 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.892722 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.892731 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.892740 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.892749 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9bs2\" (UniqueName: \"kubernetes.io/projected/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-kube-api-access-w9bs2\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.894102 4744 scope.go:117] "RemoveContainer" containerID="0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.903594 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0e74e552-fedc-4e70-ab6c-2f1bb36e2011" (UID: "0e74e552-fedc-4e70-ab6c-2f1bb36e2011"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.939494 4744 scope.go:117] "RemoveContainer" containerID="cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a" Jan 06 15:03:20 crc kubenswrapper[4744]: E0106 15:03:20.940012 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a\": container with ID starting with cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a not found: ID does not exist" containerID="cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.940056 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a"} err="failed to get container status \"cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a\": rpc error: code = NotFound desc = could not find container \"cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a\": container with ID starting with cf7abd6930369fe07636464b7fafcf763fcc833512c598b81620ccef0a8a397a not found: ID does not exist" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.940081 4744 scope.go:117] "RemoveContainer" containerID="0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef" Jan 06 15:03:20 crc kubenswrapper[4744]: E0106 15:03:20.940614 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef\": container with ID starting with 0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef not found: ID does not exist" containerID="0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.940651 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef"} err="failed to get container status \"0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef\": rpc error: code = NotFound desc = could not find container \"0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef\": container with ID starting with 0e132c9d11140c841ad8acbbc681c59641afe7701093a4fb87762738db9f2eef not found: ID does not exist" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.957433 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:03:20 crc kubenswrapper[4744]: I0106 15:03:20.995295 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e74e552-fedc-4e70-ab6c-2f1bb36e2011-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.096996 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9255f27d-3d50-47f7-92b0-ee9331a565cd-logs\") pod \"9255f27d-3d50-47f7-92b0-ee9331a565cd\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.097056 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-config-data\") pod \"9255f27d-3d50-47f7-92b0-ee9331a565cd\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.097364 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-combined-ca-bundle\") pod \"9255f27d-3d50-47f7-92b0-ee9331a565cd\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.097443 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t687x\" (UniqueName: \"kubernetes.io/projected/9255f27d-3d50-47f7-92b0-ee9331a565cd-kube-api-access-t687x\") pod \"9255f27d-3d50-47f7-92b0-ee9331a565cd\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.097467 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-nova-metadata-tls-certs\") pod \"9255f27d-3d50-47f7-92b0-ee9331a565cd\" (UID: \"9255f27d-3d50-47f7-92b0-ee9331a565cd\") " Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.099000 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9255f27d-3d50-47f7-92b0-ee9331a565cd-logs" (OuterVolumeSpecName: "logs") pod "9255f27d-3d50-47f7-92b0-ee9331a565cd" (UID: "9255f27d-3d50-47f7-92b0-ee9331a565cd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.103094 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9255f27d-3d50-47f7-92b0-ee9331a565cd-kube-api-access-t687x" (OuterVolumeSpecName: "kube-api-access-t687x") pod "9255f27d-3d50-47f7-92b0-ee9331a565cd" (UID: "9255f27d-3d50-47f7-92b0-ee9331a565cd"). InnerVolumeSpecName "kube-api-access-t687x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.129340 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9255f27d-3d50-47f7-92b0-ee9331a565cd" (UID: "9255f27d-3d50-47f7-92b0-ee9331a565cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.137090 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-config-data" (OuterVolumeSpecName: "config-data") pod "9255f27d-3d50-47f7-92b0-ee9331a565cd" (UID: "9255f27d-3d50-47f7-92b0-ee9331a565cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.158337 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "9255f27d-3d50-47f7-92b0-ee9331a565cd" (UID: "9255f27d-3d50-47f7-92b0-ee9331a565cd"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.200220 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.200252 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t687x\" (UniqueName: \"kubernetes.io/projected/9255f27d-3d50-47f7-92b0-ee9331a565cd-kube-api-access-t687x\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.200264 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.200273 4744 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9255f27d-3d50-47f7-92b0-ee9331a565cd-logs\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.200282 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9255f27d-3d50-47f7-92b0-ee9331a565cd-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.314203 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.330760 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.373549 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:21 crc kubenswrapper[4744]: E0106 15:03:21.373993 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-metadata" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.374011 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-metadata" Jan 06 15:03:21 crc kubenswrapper[4744]: E0106 15:03:21.374019 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-log" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.374027 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-log" Jan 06 15:03:21 crc kubenswrapper[4744]: E0106 15:03:21.374041 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-api" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.374047 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-api" Jan 06 15:03:21 crc kubenswrapper[4744]: E0106 15:03:21.374075 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-log" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.374082 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-log" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.374312 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-log" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.374332 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" containerName="nova-api-api" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.374347 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-metadata" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.374358 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" containerName="nova-metadata-log" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.386578 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.389774 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.394479 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.395396 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.395649 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.507924 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46daf908-4aec-483b-ba31-ee145e42de95-logs\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.508211 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqc2p\" (UniqueName: \"kubernetes.io/projected/46daf908-4aec-483b-ba31-ee145e42de95-kube-api-access-fqc2p\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.508241 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.508266 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-public-tls-certs\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.508306 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-config-data\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.508699 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-internal-tls-certs\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.610850 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46daf908-4aec-483b-ba31-ee145e42de95-logs\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.610888 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqc2p\" (UniqueName: \"kubernetes.io/projected/46daf908-4aec-483b-ba31-ee145e42de95-kube-api-access-fqc2p\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.610917 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.610937 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-public-tls-certs\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.610974 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-config-data\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.611073 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-internal-tls-certs\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.612119 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46daf908-4aec-483b-ba31-ee145e42de95-logs\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.614703 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.615517 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-public-tls-certs\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.616081 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-internal-tls-certs\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.619471 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46daf908-4aec-483b-ba31-ee145e42de95-config-data\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.635794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqc2p\" (UniqueName: \"kubernetes.io/projected/46daf908-4aec-483b-ba31-ee145e42de95-kube-api-access-fqc2p\") pod \"nova-api-0\" (UID: \"46daf908-4aec-483b-ba31-ee145e42de95\") " pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.721414 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.727442 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e74e552-fedc-4e70-ab6c-2f1bb36e2011" path="/var/lib/kubelet/pods/0e74e552-fedc-4e70-ab6c-2f1bb36e2011/volumes" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.729150 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c37294a5-5d6b-495a-a392-7744df1c0ec9" path="/var/lib/kubelet/pods/c37294a5-5d6b-495a-a392-7744df1c0ec9/volumes" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.876127 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8221b070-0d54-439f-887a-9a401d634bea","Type":"ContainerStarted","Data":"13d3897a6268795143417c0c9d2ad8ce53df46f023fdffe3e78a70cfab9a7fad"} Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.876180 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8221b070-0d54-439f-887a-9a401d634bea","Type":"ContainerStarted","Data":"1dc0dec304aaf7cd8c29856307353cf2d2778c227c9e444bc32046889b23f6ce"} Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.886507 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9255f27d-3d50-47f7-92b0-ee9331a565cd","Type":"ContainerDied","Data":"3372bc0d9f9018b84e81bab2ae44e78ef49a9a3621023291179918c6a4f14919"} Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.886563 4744 scope.go:117] "RemoveContainer" containerID="a88070ac4ba00f8e117f58818e579e3cc3c8cad215bcc3b69647a07f2884409f" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.886743 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.909766 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.909746419 podStartE2EDuration="2.909746419s" podCreationTimestamp="2026-01-06 15:03:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:03:21.896337654 +0000 UTC m=+1598.523803972" watchObservedRunningTime="2026-01-06 15:03:21.909746419 +0000 UTC m=+1598.537212737" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.927073 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.931371 4744 scope.go:117] "RemoveContainer" containerID="96d4d2f7c18c6829a3a079d4f8c558a32b22a7e809303f87b78955a94a9a568f" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.941790 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.958509 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.960446 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.967893 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.968077 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 06 15:03:21 crc kubenswrapper[4744]: I0106 15:03:21.976835 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.022540 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af5c39c-9101-4d0e-85be-a4556aef4cbb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.022598 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af5c39c-9101-4d0e-85be-a4556aef4cbb-logs\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.022715 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af5c39c-9101-4d0e-85be-a4556aef4cbb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.028435 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af5c39c-9101-4d0e-85be-a4556aef4cbb-config-data\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.028510 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfcch\" (UniqueName: \"kubernetes.io/projected/1af5c39c-9101-4d0e-85be-a4556aef4cbb-kube-api-access-qfcch\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.138556 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af5c39c-9101-4d0e-85be-a4556aef4cbb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.138634 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af5c39c-9101-4d0e-85be-a4556aef4cbb-logs\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.138975 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af5c39c-9101-4d0e-85be-a4556aef4cbb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.139116 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af5c39c-9101-4d0e-85be-a4556aef4cbb-config-data\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.139181 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1af5c39c-9101-4d0e-85be-a4556aef4cbb-logs\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.139296 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfcch\" (UniqueName: \"kubernetes.io/projected/1af5c39c-9101-4d0e-85be-a4556aef4cbb-kube-api-access-qfcch\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.143648 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1af5c39c-9101-4d0e-85be-a4556aef4cbb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.145634 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1af5c39c-9101-4d0e-85be-a4556aef4cbb-config-data\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.147090 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1af5c39c-9101-4d0e-85be-a4556aef4cbb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.154205 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfcch\" (UniqueName: \"kubernetes.io/projected/1af5c39c-9101-4d0e-85be-a4556aef4cbb-kube-api-access-qfcch\") pod \"nova-metadata-0\" (UID: \"1af5c39c-9101-4d0e-85be-a4556aef4cbb\") " pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.298428 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.304294 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 06 15:03:22 crc kubenswrapper[4744]: W0106 15:03:22.306254 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46daf908_4aec_483b_ba31_ee145e42de95.slice/crio-c661d8a9442e75280d2f651c73abc917231f94d9ee6b7272b3bfbcc1f7fe8b63 WatchSource:0}: Error finding container c661d8a9442e75280d2f651c73abc917231f94d9ee6b7272b3bfbcc1f7fe8b63: Status 404 returned error can't find the container with id c661d8a9442e75280d2f651c73abc917231f94d9ee6b7272b3bfbcc1f7fe8b63 Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.834331 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.910274 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"46daf908-4aec-483b-ba31-ee145e42de95","Type":"ContainerStarted","Data":"662bfd851b9d0d61a6b3353a57d37c225ad74f52e178268379633461d4968889"} Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.910631 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"46daf908-4aec-483b-ba31-ee145e42de95","Type":"ContainerStarted","Data":"780e93a7e4dd8dc0810b6d1ca0744c7bd2ebc2aabbff9f749ea65da562170c3a"} Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.910649 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"46daf908-4aec-483b-ba31-ee145e42de95","Type":"ContainerStarted","Data":"c661d8a9442e75280d2f651c73abc917231f94d9ee6b7272b3bfbcc1f7fe8b63"} Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.917392 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1af5c39c-9101-4d0e-85be-a4556aef4cbb","Type":"ContainerStarted","Data":"13b8f5e5927b4fb7a3781f2780ddb87240d5f313b872fef594ab3db949383f64"} Jan 06 15:03:22 crc kubenswrapper[4744]: I0106 15:03:22.942028 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.942007804 podStartE2EDuration="1.942007804s" podCreationTimestamp="2026-01-06 15:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:03:22.931700242 +0000 UTC m=+1599.559166570" watchObservedRunningTime="2026-01-06 15:03:22.942007804 +0000 UTC m=+1599.569474122" Jan 06 15:03:23 crc kubenswrapper[4744]: I0106 15:03:23.727328 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9255f27d-3d50-47f7-92b0-ee9331a565cd" path="/var/lib/kubelet/pods/9255f27d-3d50-47f7-92b0-ee9331a565cd/volumes" Jan 06 15:03:23 crc kubenswrapper[4744]: I0106 15:03:23.931194 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1af5c39c-9101-4d0e-85be-a4556aef4cbb","Type":"ContainerStarted","Data":"58634805e2a27f6bd1633badd14ac644c232983da85361a275bcabea3334f5d2"} Jan 06 15:03:23 crc kubenswrapper[4744]: I0106 15:03:23.931261 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1af5c39c-9101-4d0e-85be-a4556aef4cbb","Type":"ContainerStarted","Data":"90dca998fb98f7c7e5fe95c01da60a6657eff24bb3b04675c52d0e80f0f6e675"} Jan 06 15:03:23 crc kubenswrapper[4744]: I0106 15:03:23.958720 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9587001280000003 podStartE2EDuration="2.958700128s" podCreationTimestamp="2026-01-06 15:03:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:03:23.948880588 +0000 UTC m=+1600.576346916" watchObservedRunningTime="2026-01-06 15:03:23.958700128 +0000 UTC m=+1600.586166456" Jan 06 15:03:25 crc kubenswrapper[4744]: I0106 15:03:25.247902 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 06 15:03:27 crc kubenswrapper[4744]: I0106 15:03:27.299489 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 06 15:03:27 crc kubenswrapper[4744]: I0106 15:03:27.299760 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 06 15:03:30 crc kubenswrapper[4744]: I0106 15:03:30.248334 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 06 15:03:30 crc kubenswrapper[4744]: I0106 15:03:30.305704 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 06 15:03:31 crc kubenswrapper[4744]: I0106 15:03:31.061416 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 06 15:03:31 crc kubenswrapper[4744]: I0106 15:03:31.711318 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:03:31 crc kubenswrapper[4744]: E0106 15:03:31.711587 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:03:31 crc kubenswrapper[4744]: I0106 15:03:31.734539 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 06 15:03:31 crc kubenswrapper[4744]: I0106 15:03:31.734571 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 06 15:03:32 crc kubenswrapper[4744]: I0106 15:03:32.299386 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 06 15:03:32 crc kubenswrapper[4744]: I0106 15:03:32.299420 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 06 15:03:32 crc kubenswrapper[4744]: I0106 15:03:32.771380 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="46daf908-4aec-483b-ba31-ee145e42de95" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.9:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:03:32 crc kubenswrapper[4744]: I0106 15:03:32.771438 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="46daf908-4aec-483b-ba31-ee145e42de95" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.9:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:03:33 crc kubenswrapper[4744]: I0106 15:03:33.311314 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1af5c39c-9101-4d0e-85be-a4556aef4cbb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.10:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:03:33 crc kubenswrapper[4744]: I0106 15:03:33.311303 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1af5c39c-9101-4d0e-85be-a4556aef4cbb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.10:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.085996 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.135493 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xfw2\" (UniqueName: \"kubernetes.io/projected/70f7184f-9088-49fb-ace5-c14f5efd29e3-kube-api-access-8xfw2\") pod \"70f7184f-9088-49fb-ace5-c14f5efd29e3\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.135621 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-config-data\") pod \"70f7184f-9088-49fb-ace5-c14f5efd29e3\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.135849 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-combined-ca-bundle\") pod \"70f7184f-9088-49fb-ace5-c14f5efd29e3\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.135877 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-scripts\") pod \"70f7184f-9088-49fb-ace5-c14f5efd29e3\" (UID: \"70f7184f-9088-49fb-ace5-c14f5efd29e3\") " Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.144804 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-scripts" (OuterVolumeSpecName: "scripts") pod "70f7184f-9088-49fb-ace5-c14f5efd29e3" (UID: "70f7184f-9088-49fb-ace5-c14f5efd29e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.155601 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f7184f-9088-49fb-ace5-c14f5efd29e3-kube-api-access-8xfw2" (OuterVolumeSpecName: "kube-api-access-8xfw2") pod "70f7184f-9088-49fb-ace5-c14f5efd29e3" (UID: "70f7184f-9088-49fb-ace5-c14f5efd29e3"). InnerVolumeSpecName "kube-api-access-8xfw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.200447 4744 generic.go:334] "Generic (PLEG): container finished" podID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerID="700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827" exitCode=137 Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.200517 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerDied","Data":"700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827"} Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.200546 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"70f7184f-9088-49fb-ace5-c14f5efd29e3","Type":"ContainerDied","Data":"2759e0ed674bc145bcbdd25fda166d06f6d359f160afa69b14cc4f01e732118c"} Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.200562 4744 scope.go:117] "RemoveContainer" containerID="700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.201237 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.245252 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.245538 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xfw2\" (UniqueName: \"kubernetes.io/projected/70f7184f-9088-49fb-ace5-c14f5efd29e3-kube-api-access-8xfw2\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.300780 4744 scope.go:117] "RemoveContainer" containerID="ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.306004 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-config-data" (OuterVolumeSpecName: "config-data") pod "70f7184f-9088-49fb-ace5-c14f5efd29e3" (UID: "70f7184f-9088-49fb-ace5-c14f5efd29e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.320627 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70f7184f-9088-49fb-ace5-c14f5efd29e3" (UID: "70f7184f-9088-49fb-ace5-c14f5efd29e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.334426 4744 scope.go:117] "RemoveContainer" containerID="e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.347824 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.347855 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f7184f-9088-49fb-ace5-c14f5efd29e3-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.380206 4744 scope.go:117] "RemoveContainer" containerID="8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.412384 4744 scope.go:117] "RemoveContainer" containerID="700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827" Jan 06 15:03:41 crc kubenswrapper[4744]: E0106 15:03:41.412719 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827\": container with ID starting with 700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827 not found: ID does not exist" containerID="700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.412751 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827"} err="failed to get container status \"700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827\": rpc error: code = NotFound desc = could not find container \"700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827\": container with ID starting with 700376ac2a5808438a9833d4ba5cdae8833f2ea33d6016f22f8c81df92fd4827 not found: ID does not exist" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.412770 4744 scope.go:117] "RemoveContainer" containerID="ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595" Jan 06 15:03:41 crc kubenswrapper[4744]: E0106 15:03:41.413144 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595\": container with ID starting with ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595 not found: ID does not exist" containerID="ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.413179 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595"} err="failed to get container status \"ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595\": rpc error: code = NotFound desc = could not find container \"ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595\": container with ID starting with ecd2a533d04df2c76fcfac45143a7397c0c7c143d4ca0ca7fde06eba40319595 not found: ID does not exist" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.413198 4744 scope.go:117] "RemoveContainer" containerID="e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5" Jan 06 15:03:41 crc kubenswrapper[4744]: E0106 15:03:41.413493 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5\": container with ID starting with e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5 not found: ID does not exist" containerID="e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.413512 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5"} err="failed to get container status \"e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5\": rpc error: code = NotFound desc = could not find container \"e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5\": container with ID starting with e5f0d2d9f741d4d30cf23220699b51451059ee329273eaafb526f73ef85902b5 not found: ID does not exist" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.413523 4744 scope.go:117] "RemoveContainer" containerID="8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5" Jan 06 15:03:41 crc kubenswrapper[4744]: E0106 15:03:41.413743 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5\": container with ID starting with 8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5 not found: ID does not exist" containerID="8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.413764 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5"} err="failed to get container status \"8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5\": rpc error: code = NotFound desc = could not find container \"8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5\": container with ID starting with 8c52dc31c59a9dfeef16a3d797666fe97c4405b10f76cf0a0579ebe43b2debd5 not found: ID does not exist" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.547548 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.563584 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.583856 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Jan 06 15:03:41 crc kubenswrapper[4744]: E0106 15:03:41.584457 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-listener" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.584478 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-listener" Jan 06 15:03:41 crc kubenswrapper[4744]: E0106 15:03:41.584498 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-evaluator" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.584504 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-evaluator" Jan 06 15:03:41 crc kubenswrapper[4744]: E0106 15:03:41.584516 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-api" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.584523 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-api" Jan 06 15:03:41 crc kubenswrapper[4744]: E0106 15:03:41.584548 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-notifier" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.584554 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-notifier" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.584797 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-notifier" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.584822 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-listener" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.584840 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-evaluator" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.584852 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" containerName="aodh-api" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.604397 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.613942 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.614260 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tjbk8" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.614347 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.614645 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.614798 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.614885 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.654179 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-internal-tls-certs\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.654337 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwphb\" (UniqueName: \"kubernetes.io/projected/85034f0d-f575-44da-9f68-9d5c2c2e6f74-kube-api-access-kwphb\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.654422 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-combined-ca-bundle\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.654438 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-public-tls-certs\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.654467 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-scripts\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.654522 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-config-data\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.727208 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f7184f-9088-49fb-ace5-c14f5efd29e3" path="/var/lib/kubelet/pods/70f7184f-9088-49fb-ace5-c14f5efd29e3/volumes" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.734063 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.734464 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.736624 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.740238 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.757187 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwphb\" (UniqueName: \"kubernetes.io/projected/85034f0d-f575-44da-9f68-9d5c2c2e6f74-kube-api-access-kwphb\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.757623 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-combined-ca-bundle\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.757753 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-public-tls-certs\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.757915 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-scripts\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.758055 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-config-data\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.758197 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-internal-tls-certs\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.761442 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-scripts\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.762670 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-config-data\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.763056 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-internal-tls-certs\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.763230 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-combined-ca-bundle\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.780767 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-public-tls-certs\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:41 crc kubenswrapper[4744]: I0106 15:03:41.781543 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwphb\" (UniqueName: \"kubernetes.io/projected/85034f0d-f575-44da-9f68-9d5c2c2e6f74-kube-api-access-kwphb\") pod \"aodh-0\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " pod="openstack/aodh-0" Jan 06 15:03:42 crc kubenswrapper[4744]: I0106 15:03:42.002218 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:03:42 crc kubenswrapper[4744]: I0106 15:03:42.228431 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 06 15:03:42 crc kubenswrapper[4744]: I0106 15:03:42.248217 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 06 15:03:42 crc kubenswrapper[4744]: I0106 15:03:42.448432 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 06 15:03:42 crc kubenswrapper[4744]: I0106 15:03:42.455055 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 06 15:03:42 crc kubenswrapper[4744]: I0106 15:03:42.490690 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 06 15:03:42 crc kubenswrapper[4744]: I0106 15:03:42.496093 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 06 15:03:42 crc kubenswrapper[4744]: I0106 15:03:42.574753 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Jan 06 15:03:42 crc kubenswrapper[4744]: I0106 15:03:42.710852 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:03:42 crc kubenswrapper[4744]: E0106 15:03:42.711122 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:03:43 crc kubenswrapper[4744]: I0106 15:03:43.238468 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerStarted","Data":"54bc70450c031312f7989a99a4633cfa42fe6e015473731ac24ef8655ea017fb"} Jan 06 15:03:43 crc kubenswrapper[4744]: I0106 15:03:43.245271 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 06 15:03:44 crc kubenswrapper[4744]: I0106 15:03:44.251059 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerStarted","Data":"f6abcbfa1d6d72ebb5e38535ffbbe20cedbd190b678474b2ddc6155c7750908d"} Jan 06 15:03:45 crc kubenswrapper[4744]: I0106 15:03:45.272209 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerStarted","Data":"f4e4ab41eea7a7ea2b747213b9ef661ad9bb1c9706f988b8b509313c22fe3621"} Jan 06 15:03:46 crc kubenswrapper[4744]: I0106 15:03:46.285730 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerStarted","Data":"f6ef551a84bf8d151ee9b9d821dc1ba8fa2613e0e6488a9483a42deaedfa40f5"} Jan 06 15:03:46 crc kubenswrapper[4744]: I0106 15:03:46.286072 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerStarted","Data":"223c1c237ab1a8c64b3f384246a510c319b92d8cf96efdcaa0c4ca3a382cccd3"} Jan 06 15:03:46 crc kubenswrapper[4744]: I0106 15:03:46.313312 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.9297182670000002 podStartE2EDuration="5.313292672s" podCreationTimestamp="2026-01-06 15:03:41 +0000 UTC" firstStartedPulling="2026-01-06 15:03:42.548449989 +0000 UTC m=+1619.175916307" lastFinishedPulling="2026-01-06 15:03:45.932024394 +0000 UTC m=+1622.559490712" observedRunningTime="2026-01-06 15:03:46.308749952 +0000 UTC m=+1622.936216270" watchObservedRunningTime="2026-01-06 15:03:46.313292672 +0000 UTC m=+1622.940758990" Jan 06 15:03:54 crc kubenswrapper[4744]: I0106 15:03:54.712233 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:03:54 crc kubenswrapper[4744]: E0106 15:03:54.713443 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.024597 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-6ltjv"] Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.036581 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-6ltjv"] Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.179887 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-ptgpq"] Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.182780 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.193449 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-ptgpq"] Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.307912 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-combined-ca-bundle\") pod \"heat-db-sync-ptgpq\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.307998 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-config-data\") pod \"heat-db-sync-ptgpq\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.308149 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfsn4\" (UniqueName: \"kubernetes.io/projected/0dcc5556-fb95-4735-b08c-a11e0e2d7042-kube-api-access-vfsn4\") pod \"heat-db-sync-ptgpq\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.410412 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-combined-ca-bundle\") pod \"heat-db-sync-ptgpq\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.410517 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-config-data\") pod \"heat-db-sync-ptgpq\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.410703 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfsn4\" (UniqueName: \"kubernetes.io/projected/0dcc5556-fb95-4735-b08c-a11e0e2d7042-kube-api-access-vfsn4\") pod \"heat-db-sync-ptgpq\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.418034 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-config-data\") pod \"heat-db-sync-ptgpq\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.425886 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-combined-ca-bundle\") pod \"heat-db-sync-ptgpq\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.432331 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfsn4\" (UniqueName: \"kubernetes.io/projected/0dcc5556-fb95-4735-b08c-a11e0e2d7042-kube-api-access-vfsn4\") pod \"heat-db-sync-ptgpq\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.514575 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ptgpq" Jan 06 15:03:57 crc kubenswrapper[4744]: I0106 15:03:57.733951 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae28705-95b4-4dd8-ab04-bb1e7202ae6b" path="/var/lib/kubelet/pods/5ae28705-95b4-4dd8-ab04-bb1e7202ae6b/volumes" Jan 06 15:03:58 crc kubenswrapper[4744]: W0106 15:03:58.023383 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dcc5556_fb95_4735_b08c_a11e0e2d7042.slice/crio-d3f644f748e166b299521be819aaa2ae9b1d3176fe513c495cf5dbbe78ddfffc WatchSource:0}: Error finding container d3f644f748e166b299521be819aaa2ae9b1d3176fe513c495cf5dbbe78ddfffc: Status 404 returned error can't find the container with id d3f644f748e166b299521be819aaa2ae9b1d3176fe513c495cf5dbbe78ddfffc Jan 06 15:03:58 crc kubenswrapper[4744]: I0106 15:03:58.034187 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-ptgpq"] Jan 06 15:03:58 crc kubenswrapper[4744]: I0106 15:03:58.438271 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ptgpq" event={"ID":"0dcc5556-fb95-4735-b08c-a11e0e2d7042","Type":"ContainerStarted","Data":"d3f644f748e166b299521be819aaa2ae9b1d3176fe513c495cf5dbbe78ddfffc"} Jan 06 15:03:59 crc kubenswrapper[4744]: I0106 15:03:59.391359 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:03:59 crc kubenswrapper[4744]: I0106 15:03:59.391754 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="ceilometer-central-agent" containerID="cri-o://f9fda546f2326b3980c08b17525d568a5f2c7fb9aa7958d4773c7010644a0d7e" gracePeriod=30 Jan 06 15:03:59 crc kubenswrapper[4744]: I0106 15:03:59.391771 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="proxy-httpd" containerID="cri-o://481493d2e869b5d33bc577565682801a87433163305e82bb60062f4559087d28" gracePeriod=30 Jan 06 15:03:59 crc kubenswrapper[4744]: I0106 15:03:59.391831 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="ceilometer-notification-agent" containerID="cri-o://7555282a4e1ae18eaf7b74f6e033056f94245ebf901084f171fedc45d006238a" gracePeriod=30 Jan 06 15:03:59 crc kubenswrapper[4744]: I0106 15:03:59.391870 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="sg-core" containerID="cri-o://ceea7c24813d0da0f03e90e5acfd4247ca59fed292baa7666a8555ee19da01cb" gracePeriod=30 Jan 06 15:03:59 crc kubenswrapper[4744]: I0106 15:03:59.893682 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Jan 06 15:04:00 crc kubenswrapper[4744]: I0106 15:04:00.467213 4744 generic.go:334] "Generic (PLEG): container finished" podID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerID="481493d2e869b5d33bc577565682801a87433163305e82bb60062f4559087d28" exitCode=0 Jan 06 15:04:00 crc kubenswrapper[4744]: I0106 15:04:00.467572 4744 generic.go:334] "Generic (PLEG): container finished" podID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerID="ceea7c24813d0da0f03e90e5acfd4247ca59fed292baa7666a8555ee19da01cb" exitCode=2 Jan 06 15:04:00 crc kubenswrapper[4744]: I0106 15:04:00.467592 4744 generic.go:334] "Generic (PLEG): container finished" podID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerID="f9fda546f2326b3980c08b17525d568a5f2c7fb9aa7958d4773c7010644a0d7e" exitCode=0 Jan 06 15:04:00 crc kubenswrapper[4744]: I0106 15:04:00.467284 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerDied","Data":"481493d2e869b5d33bc577565682801a87433163305e82bb60062f4559087d28"} Jan 06 15:04:00 crc kubenswrapper[4744]: I0106 15:04:00.467638 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerDied","Data":"ceea7c24813d0da0f03e90e5acfd4247ca59fed292baa7666a8555ee19da01cb"} Jan 06 15:04:00 crc kubenswrapper[4744]: I0106 15:04:00.467655 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerDied","Data":"f9fda546f2326b3980c08b17525d568a5f2c7fb9aa7958d4773c7010644a0d7e"} Jan 06 15:04:00 crc kubenswrapper[4744]: I0106 15:04:00.948607 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.528463 4744 generic.go:334] "Generic (PLEG): container finished" podID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerID="7555282a4e1ae18eaf7b74f6e033056f94245ebf901084f171fedc45d006238a" exitCode=0 Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.528944 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerDied","Data":"7555282a4e1ae18eaf7b74f6e033056f94245ebf901084f171fedc45d006238a"} Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.653408 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.797443 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rx8w\" (UniqueName: \"kubernetes.io/projected/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-kube-api-access-2rx8w\") pod \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.797611 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-combined-ca-bundle\") pod \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.797657 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-sg-core-conf-yaml\") pod \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.797710 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-scripts\") pod \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.797769 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-run-httpd\") pod \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.797794 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-log-httpd\") pod \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.797829 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-config-data\") pod \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.797906 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-ceilometer-tls-certs\") pod \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\" (UID: \"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49\") " Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.799120 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" (UID: "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.799622 4744 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.800791 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" (UID: "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.820513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-kube-api-access-2rx8w" (OuterVolumeSpecName: "kube-api-access-2rx8w") pod "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" (UID: "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49"). InnerVolumeSpecName "kube-api-access-2rx8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.820519 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-scripts" (OuterVolumeSpecName: "scripts") pod "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" (UID: "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.884026 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" (UID: "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.903360 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rx8w\" (UniqueName: \"kubernetes.io/projected/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-kube-api-access-2rx8w\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.903396 4744 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.903407 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.903416 4744 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.914042 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" (UID: "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.977419 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" (UID: "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:04:03 crc kubenswrapper[4744]: I0106 15:04:03.997897 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-config-data" (OuterVolumeSpecName: "config-data") pod "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" (UID: "1e54a0df-e1ed-44e0-8278-2c9bb03e5d49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.005090 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.005115 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.005124 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.300168 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-2" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="rabbitmq" containerID="cri-o://595b0c188fa47a16c899d11b4a1a07707b58c92ed185b449bf699f6384597f4e" gracePeriod=604796 Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.535992 4744 scope.go:117] "RemoveContainer" containerID="c7766c02ae0bac5d27289481b8d90064c15e4b93c06f99ef0a12a9699f8225a2" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.545152 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1e54a0df-e1ed-44e0-8278-2c9bb03e5d49","Type":"ContainerDied","Data":"3b51beadcd1518d0e7fc3ff6b4e556268cfda31dde8079af44abfe50eba438af"} Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.545226 4744 scope.go:117] "RemoveContainer" containerID="481493d2e869b5d33bc577565682801a87433163305e82bb60062f4559087d28" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.545407 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.626297 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.637240 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.682105 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:04:04 crc kubenswrapper[4744]: E0106 15:04:04.682740 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="ceilometer-notification-agent" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.682764 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="ceilometer-notification-agent" Jan 06 15:04:04 crc kubenswrapper[4744]: E0106 15:04:04.682806 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="ceilometer-central-agent" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.682815 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="ceilometer-central-agent" Jan 06 15:04:04 crc kubenswrapper[4744]: E0106 15:04:04.682838 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="sg-core" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.682846 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="sg-core" Jan 06 15:04:04 crc kubenswrapper[4744]: E0106 15:04:04.682876 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="proxy-httpd" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.682884 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="proxy-httpd" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.683185 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="proxy-httpd" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.683205 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="ceilometer-central-agent" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.683226 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="ceilometer-notification-agent" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.683261 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" containerName="sg-core" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.686312 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.689613 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.689801 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.695946 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.697636 4744 scope.go:117] "RemoveContainer" containerID="55c7ebd388c47561fc4ecc894a9ba12f3b56af6569453cf4d560ae54646623b3" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.698842 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.813291 4744 scope.go:117] "RemoveContainer" containerID="ceea7c24813d0da0f03e90e5acfd4247ca59fed292baa7666a8555ee19da01cb" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.826330 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-config-data\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.826375 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32ce34b0-736e-4d20-9c3b-4c76656092dc-run-httpd\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.826450 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.826553 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgcbv\" (UniqueName: \"kubernetes.io/projected/32ce34b0-736e-4d20-9c3b-4c76656092dc-kube-api-access-cgcbv\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.826594 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.826693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-scripts\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.826834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32ce34b0-736e-4d20-9c3b-4c76656092dc-log-httpd\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.826964 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.929996 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-scripts\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.930062 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32ce34b0-736e-4d20-9c3b-4c76656092dc-log-httpd\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.930109 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.930205 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-config-data\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.930249 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32ce34b0-736e-4d20-9c3b-4c76656092dc-run-httpd\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.930314 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.930395 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgcbv\" (UniqueName: \"kubernetes.io/projected/32ce34b0-736e-4d20-9c3b-4c76656092dc-kube-api-access-cgcbv\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.930437 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.933588 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32ce34b0-736e-4d20-9c3b-4c76656092dc-log-httpd\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.934227 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/32ce34b0-736e-4d20-9c3b-4c76656092dc-run-httpd\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.938023 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-config-data\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.938306 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-scripts\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.942678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.943472 4744 scope.go:117] "RemoveContainer" containerID="7555282a4e1ae18eaf7b74f6e033056f94245ebf901084f171fedc45d006238a" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.950943 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgcbv\" (UniqueName: \"kubernetes.io/projected/32ce34b0-736e-4d20-9c3b-4c76656092dc-kube-api-access-cgcbv\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.964611 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:04 crc kubenswrapper[4744]: I0106 15:04:04.967501 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/32ce34b0-736e-4d20-9c3b-4c76656092dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"32ce34b0-736e-4d20-9c3b-4c76656092dc\") " pod="openstack/ceilometer-0" Jan 06 15:04:05 crc kubenswrapper[4744]: I0106 15:04:05.021143 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 06 15:04:05 crc kubenswrapper[4744]: I0106 15:04:05.218678 4744 scope.go:117] "RemoveContainer" containerID="f9fda546f2326b3980c08b17525d568a5f2c7fb9aa7958d4773c7010644a0d7e" Jan 06 15:04:05 crc kubenswrapper[4744]: I0106 15:04:05.387776 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerName="rabbitmq" containerID="cri-o://e8e69d68f403efbf1a5c144e3e5408e06f7ea892efbde73e923d23e93353ab57" gracePeriod=604796 Jan 06 15:04:05 crc kubenswrapper[4744]: I0106 15:04:05.577774 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 06 15:04:05 crc kubenswrapper[4744]: W0106 15:04:05.587666 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32ce34b0_736e_4d20_9c3b_4c76656092dc.slice/crio-4f397958665b35984c3bcc6fa15e123546f8d3c8c93af353f7739a6e3300846d WatchSource:0}: Error finding container 4f397958665b35984c3bcc6fa15e123546f8d3c8c93af353f7739a6e3300846d: Status 404 returned error can't find the container with id 4f397958665b35984c3bcc6fa15e123546f8d3c8c93af353f7739a6e3300846d Jan 06 15:04:05 crc kubenswrapper[4744]: I0106 15:04:05.734629 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e54a0df-e1ed-44e0-8278-2c9bb03e5d49" path="/var/lib/kubelet/pods/1e54a0df-e1ed-44e0-8278-2c9bb03e5d49/volumes" Jan 06 15:04:06 crc kubenswrapper[4744]: I0106 15:04:06.633774 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32ce34b0-736e-4d20-9c3b-4c76656092dc","Type":"ContainerStarted","Data":"4f397958665b35984c3bcc6fa15e123546f8d3c8c93af353f7739a6e3300846d"} Jan 06 15:04:06 crc kubenswrapper[4744]: I0106 15:04:06.711177 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:04:06 crc kubenswrapper[4744]: E0106 15:04:06.711566 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:04:08 crc kubenswrapper[4744]: I0106 15:04:08.720839 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Jan 06 15:04:09 crc kubenswrapper[4744]: I0106 15:04:09.048592 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Jan 06 15:04:12 crc kubenswrapper[4744]: I0106 15:04:12.713712 4744 generic.go:334] "Generic (PLEG): container finished" podID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerID="e8e69d68f403efbf1a5c144e3e5408e06f7ea892efbde73e923d23e93353ab57" exitCode=0 Jan 06 15:04:12 crc kubenswrapper[4744]: I0106 15:04:12.713778 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4","Type":"ContainerDied","Data":"e8e69d68f403efbf1a5c144e3e5408e06f7ea892efbde73e923d23e93353ab57"} Jan 06 15:04:12 crc kubenswrapper[4744]: I0106 15:04:12.716943 4744 generic.go:334] "Generic (PLEG): container finished" podID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerID="595b0c188fa47a16c899d11b4a1a07707b58c92ed185b449bf699f6384597f4e" exitCode=0 Jan 06 15:04:12 crc kubenswrapper[4744]: I0106 15:04:12.716981 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"d030be7f-cfb2-4335-a2f8-7da8f7430b17","Type":"ContainerDied","Data":"595b0c188fa47a16c899d11b4a1a07707b58c92ed185b449bf699f6384597f4e"} Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.463354 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.622203 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-config-data\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.622271 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-plugins-conf\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.622331 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-erlang-cookie-secret\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.622399 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rd5p\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-kube-api-access-9rd5p\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.622428 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-tls\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.622897 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.622940 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-confd\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.622997 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-plugins\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.623040 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-erlang-cookie\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.623087 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-server-conf\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.623087 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.623376 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-pod-info\") pod \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\" (UID: \"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4\") " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.623864 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.624145 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.624184 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.624527 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.633293 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.644870 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-pod-info" (OuterVolumeSpecName: "pod-info") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.648841 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.649882 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-kube-api-access-9rd5p" (OuterVolumeSpecName: "kube-api-access-9rd5p") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "kube-api-access-9rd5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.669119 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-config-data" (OuterVolumeSpecName: "config-data") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.677065 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a" (OuterVolumeSpecName: "persistence") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.727417 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") on node \"crc\" " Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.727449 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.727459 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-pod-info\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.727469 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.727481 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.727490 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rd5p\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-kube-api-access-9rd5p\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.727498 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.738505 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-server-conf" (OuterVolumeSpecName: "server-conf") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.777621 4744 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.777753 4744 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a") on node "crc" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.819297 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" (UID: "f1c00f21-cf61-4eb7-ac82-0c5bae9bace4"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.820743 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f1c00f21-cf61-4eb7-ac82-0c5bae9bace4","Type":"ContainerDied","Data":"d48c7f4b6fbbd43e4b4c36ab9bc02c0d6374fbf65fd6006c796b98ea5659dc90"} Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.820791 4744 scope.go:117] "RemoveContainer" containerID="e8e69d68f403efbf1a5c144e3e5408e06f7ea892efbde73e923d23e93353ab57" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.820808 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.829642 4744 reconciler_common.go:293] "Volume detached for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.829668 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.829679 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4-server-conf\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.922811 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.949611 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.982225 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 06 15:04:14 crc kubenswrapper[4744]: E0106 15:04:14.982787 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerName="setup-container" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.982802 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerName="setup-container" Jan 06 15:04:14 crc kubenswrapper[4744]: E0106 15:04:14.982854 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerName="rabbitmq" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.982860 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerName="rabbitmq" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.983060 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" containerName="rabbitmq" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.984807 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.989230 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.989473 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.989648 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.989995 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.990261 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.990615 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gsssz" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.990832 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 06 15:04:14 crc kubenswrapper[4744]: I0106 15:04:14.998600 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034307 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034380 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034412 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp5n6\" (UniqueName: \"kubernetes.io/projected/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-kube-api-access-cp5n6\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034451 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034569 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034596 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034615 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034705 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034723 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034749 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.034992 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.139807 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp5n6\" (UniqueName: \"kubernetes.io/projected/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-kube-api-access-cp5n6\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.139916 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.140017 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.140043 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.140070 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.140197 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.140223 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.140257 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.140282 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.140319 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.140353 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.142644 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.142916 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.146021 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.146237 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.146625 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.149545 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.150293 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.150382 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.150570 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.152252 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.152314 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/86ab0050a515986cbb8c4600c50553d9499b965b1a548dd8be0d02f2d54017c7/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:15 crc kubenswrapper[4744]: I0106 15:04:15.166299 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp5n6\" (UniqueName: \"kubernetes.io/projected/a33bb3f6-57ba-4e75-af0b-f9895afb79aa-kube-api-access-cp5n6\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:16 crc kubenswrapper[4744]: I0106 15:04:16.150826 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1c00f21-cf61-4eb7-ac82-0c5bae9bace4" path="/var/lib/kubelet/pods/f1c00f21-cf61-4eb7-ac82-0c5bae9bace4/volumes" Jan 06 15:04:16 crc kubenswrapper[4744]: I0106 15:04:16.385371 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fcdc5ebe-ec41-4b0b-9e47-21a568614e5a\") pod \"rabbitmq-cell1-server-0\" (UID: \"a33bb3f6-57ba-4e75-af0b-f9895afb79aa\") " pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:16 crc kubenswrapper[4744]: I0106 15:04:16.518930 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.257313 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-tdwsz"] Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.260198 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.276583 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.295641 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-tdwsz"] Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.338680 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.338744 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.338835 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.338860 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-svc\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.338885 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.338966 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-config\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.338997 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgzgw\" (UniqueName: \"kubernetes.io/projected/a931a95f-eaca-43d5-93ed-3d66ade2f347-kube-api-access-dgzgw\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.443218 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.443291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.443362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.443391 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-svc\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.443414 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.443456 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-config\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.443485 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgzgw\" (UniqueName: \"kubernetes.io/projected/a931a95f-eaca-43d5-93ed-3d66ade2f347-kube-api-access-dgzgw\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.444339 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-swift-storage-0\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.444645 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-sb\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.444976 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-svc\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.444995 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-openstack-edpm-ipam\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.445065 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-nb\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.445419 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-config\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.463931 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgzgw\" (UniqueName: \"kubernetes.io/projected/a931a95f-eaca-43d5-93ed-3d66ade2f347-kube-api-access-dgzgw\") pod \"dnsmasq-dns-594cb89c79-tdwsz\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:17 crc kubenswrapper[4744]: I0106 15:04:17.604562 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:21 crc kubenswrapper[4744]: I0106 15:04:21.710890 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:04:21 crc kubenswrapper[4744]: E0106 15:04:21.711668 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:04:23 crc kubenswrapper[4744]: I0106 15:04:23.727016 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: i/o timeout" Jan 06 15:04:26 crc kubenswrapper[4744]: I0106 15:04:26.997278 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.006141 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"d030be7f-cfb2-4335-a2f8-7da8f7430b17","Type":"ContainerDied","Data":"5496c5c0eb0ceb1a920069e3013be3ed96f77de0f5bd779d2316d842b284844c"} Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.149411 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-config-data\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.149698 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d030be7f-cfb2-4335-a2f8-7da8f7430b17-pod-info\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.149769 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-confd\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.149808 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-plugins-conf\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.149851 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-erlang-cookie\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.149931 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-server-conf\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.149985 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbxqc\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-kube-api-access-wbxqc\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.150021 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-plugins\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.150100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-tls\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.150751 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.150797 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d030be7f-cfb2-4335-a2f8-7da8f7430b17-erlang-cookie-secret\") pod \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\" (UID: \"d030be7f-cfb2-4335-a2f8-7da8f7430b17\") " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.152318 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.152776 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.159285 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d030be7f-cfb2-4335-a2f8-7da8f7430b17-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.159982 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.160354 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-kube-api-access-wbxqc" (OuterVolumeSpecName: "kube-api-access-wbxqc") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "kube-api-access-wbxqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.181517 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d030be7f-cfb2-4335-a2f8-7da8f7430b17-pod-info" (OuterVolumeSpecName: "pod-info") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.184484 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.198507 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9" (OuterVolumeSpecName: "persistence") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.200666 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-config-data" (OuterVolumeSpecName: "config-data") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.252181 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-server-conf" (OuterVolumeSpecName: "server-conf") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253807 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253824 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253835 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-server-conf\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253844 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbxqc\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-kube-api-access-wbxqc\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253852 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253860 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253889 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") on node \"crc\" " Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253901 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d030be7f-cfb2-4335-a2f8-7da8f7430b17-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253910 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d030be7f-cfb2-4335-a2f8-7da8f7430b17-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.253917 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d030be7f-cfb2-4335-a2f8-7da8f7430b17-pod-info\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.306053 4744 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.306227 4744 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9") on node "crc" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.355887 4744 reconciler_common.go:293] "Volume detached for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.371660 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d030be7f-cfb2-4335-a2f8-7da8f7430b17" (UID: "d030be7f-cfb2-4335-a2f8-7da8f7430b17"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:27 crc kubenswrapper[4744]: I0106 15:04:27.458842 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d030be7f-cfb2-4335-a2f8-7da8f7430b17-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:27 crc kubenswrapper[4744]: E0106 15:04:27.617902 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Jan 06 15:04:27 crc kubenswrapper[4744]: E0106 15:04:27.617971 4744 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Jan 06 15:04:27 crc kubenswrapper[4744]: E0106 15:04:27.618121 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nch546h55fh6hc7h589h599h658h65h96h555h666hdch5d5h65chdh98h5d6h5fhch59bh5c5h576h5c4hf7hfch544hd5hd6hbbhffh64bq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cgcbv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(32ce34b0-736e-4d20-9c3b-4c76656092dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.043126 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.079239 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.096367 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-2"] Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.132906 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Jan 06 15:04:28 crc kubenswrapper[4744]: E0106 15:04:28.133460 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="rabbitmq" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.133479 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="rabbitmq" Jan 06 15:04:28 crc kubenswrapper[4744]: E0106 15:04:28.133508 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="setup-container" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.133515 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="setup-container" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.133781 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" containerName="rabbitmq" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.134982 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: E0106 15:04:28.159327 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Jan 06 15:04:28 crc kubenswrapper[4744]: E0106 15:04:28.159389 4744 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Jan 06 15:04:28 crc kubenswrapper[4744]: E0106 15:04:28.159506 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vfsn4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-ptgpq_openstack(0dcc5556-fb95-4735-b08c-a11e0e2d7042): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 15:04:28 crc kubenswrapper[4744]: E0106 15:04:28.160597 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-ptgpq" podUID="0dcc5556-fb95-4735-b08c-a11e0e2d7042" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.180478 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.187626 4744 scope.go:117] "RemoveContainer" containerID="23858f13f4cc753d4995c5de460b682039086270f2029befd5f5afb9b8976ea7" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.283667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.283731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.283773 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-pod-info\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.283792 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.283826 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cshs\" (UniqueName: \"kubernetes.io/projected/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-kube-api-access-2cshs\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.283916 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.283932 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-config-data\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.284010 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-server-conf\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.284080 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.284109 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.284141 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.366001 4744 scope.go:117] "RemoveContainer" containerID="595b0c188fa47a16c899d11b4a1a07707b58c92ed185b449bf699f6384597f4e" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.387944 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.390400 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.390484 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-config-data\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.390527 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-server-conf\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.390693 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.390736 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.390775 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.390851 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.390925 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.391003 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.391026 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-pod-info\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.391100 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cshs\" (UniqueName: \"kubernetes.io/projected/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-kube-api-access-2cshs\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.391244 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-config-data\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.391794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.391863 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-server-conf\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.392395 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.399778 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.399849 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.400126 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-pod-info\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.403939 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.403970 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fab1f2299b47ba0d1659e05dfaf91881496baf7a2af79f5e175e50cd9ff7b4c7/globalmount\"" pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.404501 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.414049 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cshs\" (UniqueName: \"kubernetes.io/projected/b4c3d9bb-3170-4372-8cdd-ec1bfe092237-kube-api-access-2cshs\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.418327 4744 scope.go:117] "RemoveContainer" containerID="7bdc826178224ec8f48dc20281097da9db5ad81ce52b8030850b40e45948d25b" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.498937 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e9f7ab89-d132-4d55-81c0-ec7e3a1324a9\") pod \"rabbitmq-server-2\" (UID: \"b4c3d9bb-3170-4372-8cdd-ec1bfe092237\") " pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.776161 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.889834 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-tdwsz"] Jan 06 15:04:28 crc kubenswrapper[4744]: I0106 15:04:28.947337 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 06 15:04:28 crc kubenswrapper[4744]: W0106 15:04:28.949986 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda931a95f_eaca_43d5_93ed_3d66ade2f347.slice/crio-30044594a919137732b849710d82397327544e29babed63976fd961d083b65ea WatchSource:0}: Error finding container 30044594a919137732b849710d82397327544e29babed63976fd961d083b65ea: Status 404 returned error can't find the container with id 30044594a919137732b849710d82397327544e29babed63976fd961d083b65ea Jan 06 15:04:28 crc kubenswrapper[4744]: W0106 15:04:28.954365 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda33bb3f6_57ba_4e75_af0b_f9895afb79aa.slice/crio-cfedda3da74d95cb33571e6aac9f5eb3635f473b40e9d0e20d8ff2188151cfc7 WatchSource:0}: Error finding container cfedda3da74d95cb33571e6aac9f5eb3635f473b40e9d0e20d8ff2188151cfc7: Status 404 returned error can't find the container with id cfedda3da74d95cb33571e6aac9f5eb3635f473b40e9d0e20d8ff2188151cfc7 Jan 06 15:04:29 crc kubenswrapper[4744]: I0106 15:04:29.064786 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32ce34b0-736e-4d20-9c3b-4c76656092dc","Type":"ContainerStarted","Data":"5c94da290beb4a52b6844a504a41f5d30178034a487af385ffb3da93df0228d8"} Jan 06 15:04:29 crc kubenswrapper[4744]: I0106 15:04:29.065700 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" event={"ID":"a931a95f-eaca-43d5-93ed-3d66ade2f347","Type":"ContainerStarted","Data":"30044594a919137732b849710d82397327544e29babed63976fd961d083b65ea"} Jan 06 15:04:29 crc kubenswrapper[4744]: I0106 15:04:29.068126 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a33bb3f6-57ba-4e75-af0b-f9895afb79aa","Type":"ContainerStarted","Data":"cfedda3da74d95cb33571e6aac9f5eb3635f473b40e9d0e20d8ff2188151cfc7"} Jan 06 15:04:29 crc kubenswrapper[4744]: E0106 15:04:29.073175 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-ptgpq" podUID="0dcc5556-fb95-4735-b08c-a11e0e2d7042" Jan 06 15:04:29 crc kubenswrapper[4744]: I0106 15:04:29.329904 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Jan 06 15:04:29 crc kubenswrapper[4744]: W0106 15:04:29.339189 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4c3d9bb_3170_4372_8cdd_ec1bfe092237.slice/crio-e7202c6ea1587b3e686d3ae13d178a57206028eae069da63dd88359a4ab53fdc WatchSource:0}: Error finding container e7202c6ea1587b3e686d3ae13d178a57206028eae069da63dd88359a4ab53fdc: Status 404 returned error can't find the container with id e7202c6ea1587b3e686d3ae13d178a57206028eae069da63dd88359a4ab53fdc Jan 06 15:04:29 crc kubenswrapper[4744]: I0106 15:04:29.741111 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d030be7f-cfb2-4335-a2f8-7da8f7430b17" path="/var/lib/kubelet/pods/d030be7f-cfb2-4335-a2f8-7da8f7430b17/volumes" Jan 06 15:04:30 crc kubenswrapper[4744]: I0106 15:04:30.081420 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"b4c3d9bb-3170-4372-8cdd-ec1bfe092237","Type":"ContainerStarted","Data":"e7202c6ea1587b3e686d3ae13d178a57206028eae069da63dd88359a4ab53fdc"} Jan 06 15:04:30 crc kubenswrapper[4744]: I0106 15:04:30.084654 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32ce34b0-736e-4d20-9c3b-4c76656092dc","Type":"ContainerStarted","Data":"a8931e91e2391ad7cbd514c219b0189e19a5b17054df450893bff57b82751e65"} Jan 06 15:04:30 crc kubenswrapper[4744]: I0106 15:04:30.086721 4744 generic.go:334] "Generic (PLEG): container finished" podID="a931a95f-eaca-43d5-93ed-3d66ade2f347" containerID="f9ff112feeafafd32e74ee1f6fee7dbe0710eff3ea7d9f186164339406308e1c" exitCode=0 Jan 06 15:04:30 crc kubenswrapper[4744]: I0106 15:04:30.086745 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" event={"ID":"a931a95f-eaca-43d5-93ed-3d66ade2f347","Type":"ContainerDied","Data":"f9ff112feeafafd32e74ee1f6fee7dbe0710eff3ea7d9f186164339406308e1c"} Jan 06 15:04:31 crc kubenswrapper[4744]: I0106 15:04:31.109417 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a33bb3f6-57ba-4e75-af0b-f9895afb79aa","Type":"ContainerStarted","Data":"ad125241967b0f11778000d67b4ebb9e7b4aadd6bfcd5632e40e41e1882dd4fd"} Jan 06 15:04:31 crc kubenswrapper[4744]: E0106 15:04:31.793626 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="32ce34b0-736e-4d20-9c3b-4c76656092dc" Jan 06 15:04:32 crc kubenswrapper[4744]: I0106 15:04:32.123594 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" event={"ID":"a931a95f-eaca-43d5-93ed-3d66ade2f347","Type":"ContainerStarted","Data":"bd4bc99d5a0c08d8c164555b9afb38f0bb04c9142db681d0a143da501a3f7805"} Jan 06 15:04:32 crc kubenswrapper[4744]: I0106 15:04:32.123762 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:32 crc kubenswrapper[4744]: I0106 15:04:32.125986 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"b4c3d9bb-3170-4372-8cdd-ec1bfe092237","Type":"ContainerStarted","Data":"052672837a8e2c73135ae5ac6f5cb308335022fab3b7fc9052df9a1db50a9e61"} Jan 06 15:04:32 crc kubenswrapper[4744]: I0106 15:04:32.128415 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32ce34b0-736e-4d20-9c3b-4c76656092dc","Type":"ContainerStarted","Data":"6531f89601e5accf3c4a78896208385333d0c70dc238ed529da4555d1fb47921"} Jan 06 15:04:32 crc kubenswrapper[4744]: I0106 15:04:32.129452 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 06 15:04:32 crc kubenswrapper[4744]: E0106 15:04:32.131519 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="32ce34b0-736e-4d20-9c3b-4c76656092dc" Jan 06 15:04:32 crc kubenswrapper[4744]: I0106 15:04:32.178220 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" podStartSLOduration=15.178192265 podStartE2EDuration="15.178192265s" podCreationTimestamp="2026-01-06 15:04:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:04:32.161776521 +0000 UTC m=+1668.789242859" watchObservedRunningTime="2026-01-06 15:04:32.178192265 +0000 UTC m=+1668.805658583" Jan 06 15:04:32 crc kubenswrapper[4744]: I0106 15:04:32.712079 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:04:32 crc kubenswrapper[4744]: E0106 15:04:32.712838 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:04:33 crc kubenswrapper[4744]: E0106 15:04:33.142879 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="32ce34b0-736e-4d20-9c3b-4c76656092dc" Jan 06 15:04:37 crc kubenswrapper[4744]: I0106 15:04:37.606470 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:37 crc kubenswrapper[4744]: I0106 15:04:37.675064 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-x9rf9"] Jan 06 15:04:37 crc kubenswrapper[4744]: I0106 15:04:37.675364 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" podUID="7405beaf-33c5-493b-bba7-83419a434632" containerName="dnsmasq-dns" containerID="cri-o://e3637667911620da1180a235994b6954085165329abea4bb7a1f23fb84667847" gracePeriod=10 Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.847642 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-nb6pm"] Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.850469 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.861485 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-nb6pm"] Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.977148 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.977236 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.977262 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.977754 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.977874 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-config\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.977969 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:38 crc kubenswrapper[4744]: I0106 15:04:38.978234 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkzlp\" (UniqueName: \"kubernetes.io/projected/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-kube-api-access-bkzlp\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.081075 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkzlp\" (UniqueName: \"kubernetes.io/projected/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-kube-api-access-bkzlp\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.081182 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.081232 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.081254 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.081362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.081393 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-config\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.081422 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.082402 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-dns-swift-storage-0\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.082436 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-dns-svc\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.082444 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-ovsdbserver-sb\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.082533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-openstack-edpm-ipam\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.083091 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-ovsdbserver-nb\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.084287 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-config\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.102665 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkzlp\" (UniqueName: \"kubernetes.io/projected/aa8be235-0419-4ff3-8e76-d48ddbcfe47f-kube-api-access-bkzlp\") pod \"dnsmasq-dns-5596c69fcc-nb6pm\" (UID: \"aa8be235-0419-4ff3-8e76-d48ddbcfe47f\") " pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.172460 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.216989 4744 generic.go:334] "Generic (PLEG): container finished" podID="7405beaf-33c5-493b-bba7-83419a434632" containerID="e3637667911620da1180a235994b6954085165329abea4bb7a1f23fb84667847" exitCode=0 Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.217039 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" event={"ID":"7405beaf-33c5-493b-bba7-83419a434632","Type":"ContainerDied","Data":"e3637667911620da1180a235994b6954085165329abea4bb7a1f23fb84667847"} Jan 06 15:04:39 crc kubenswrapper[4744]: I0106 15:04:39.689615 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5596c69fcc-nb6pm"] Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.231823 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" event={"ID":"7405beaf-33c5-493b-bba7-83419a434632","Type":"ContainerDied","Data":"a3d9e853b4458da1315369134ae769779b23055e524692501caed53c886918fa"} Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.232861 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3d9e853b4458da1315369134ae769779b23055e524692501caed53c886918fa" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.234384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" event={"ID":"aa8be235-0419-4ff3-8e76-d48ddbcfe47f","Type":"ContainerStarted","Data":"c061019dcf01578c6546145f58c3f1d0ddb0cbfef9d4e531c160790ca6d4804e"} Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.248527 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.315590 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-sb\") pod \"7405beaf-33c5-493b-bba7-83419a434632\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.316220 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-nb\") pod \"7405beaf-33c5-493b-bba7-83419a434632\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.316327 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-svc\") pod \"7405beaf-33c5-493b-bba7-83419a434632\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.316451 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-config\") pod \"7405beaf-33c5-493b-bba7-83419a434632\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.316542 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-swift-storage-0\") pod \"7405beaf-33c5-493b-bba7-83419a434632\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.316630 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6lmt\" (UniqueName: \"kubernetes.io/projected/7405beaf-33c5-493b-bba7-83419a434632-kube-api-access-r6lmt\") pod \"7405beaf-33c5-493b-bba7-83419a434632\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.347848 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7405beaf-33c5-493b-bba7-83419a434632-kube-api-access-r6lmt" (OuterVolumeSpecName: "kube-api-access-r6lmt") pod "7405beaf-33c5-493b-bba7-83419a434632" (UID: "7405beaf-33c5-493b-bba7-83419a434632"). InnerVolumeSpecName "kube-api-access-r6lmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.415768 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7405beaf-33c5-493b-bba7-83419a434632" (UID: "7405beaf-33c5-493b-bba7-83419a434632"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.418833 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-config" (OuterVolumeSpecName: "config") pod "7405beaf-33c5-493b-bba7-83419a434632" (UID: "7405beaf-33c5-493b-bba7-83419a434632"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.419102 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-config\") pod \"7405beaf-33c5-493b-bba7-83419a434632\" (UID: \"7405beaf-33c5-493b-bba7-83419a434632\") " Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.419770 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.419793 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6lmt\" (UniqueName: \"kubernetes.io/projected/7405beaf-33c5-493b-bba7-83419a434632-kube-api-access-r6lmt\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:40 crc kubenswrapper[4744]: W0106 15:04:40.419870 4744 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/7405beaf-33c5-493b-bba7-83419a434632/volumes/kubernetes.io~configmap/config Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.419881 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-config" (OuterVolumeSpecName: "config") pod "7405beaf-33c5-493b-bba7-83419a434632" (UID: "7405beaf-33c5-493b-bba7-83419a434632"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.427753 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7405beaf-33c5-493b-bba7-83419a434632" (UID: "7405beaf-33c5-493b-bba7-83419a434632"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.447293 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7405beaf-33c5-493b-bba7-83419a434632" (UID: "7405beaf-33c5-493b-bba7-83419a434632"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.476462 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7405beaf-33c5-493b-bba7-83419a434632" (UID: "7405beaf-33c5-493b-bba7-83419a434632"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.522107 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.522140 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.522152 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:40 crc kubenswrapper[4744]: I0106 15:04:40.522173 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7405beaf-33c5-493b-bba7-83419a434632-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:41 crc kubenswrapper[4744]: I0106 15:04:41.254304 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" event={"ID":"aa8be235-0419-4ff3-8e76-d48ddbcfe47f","Type":"ContainerStarted","Data":"68c649576468951da258319abbce985baf1ed73a15d779d1e7b300baf1d3409a"} Jan 06 15:04:41 crc kubenswrapper[4744]: I0106 15:04:41.254324 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d99f6bc7f-x9rf9" Jan 06 15:04:41 crc kubenswrapper[4744]: I0106 15:04:41.307387 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-x9rf9"] Jan 06 15:04:41 crc kubenswrapper[4744]: I0106 15:04:41.323158 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d99f6bc7f-x9rf9"] Jan 06 15:04:41 crc kubenswrapper[4744]: E0106 15:04:41.551806 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7405beaf_33c5_493b_bba7_83419a434632.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7405beaf_33c5_493b_bba7_83419a434632.slice/crio-a3d9e853b4458da1315369134ae769779b23055e524692501caed53c886918fa\": RecentStats: unable to find data in memory cache]" Jan 06 15:04:41 crc kubenswrapper[4744]: I0106 15:04:41.727301 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7405beaf-33c5-493b-bba7-83419a434632" path="/var/lib/kubelet/pods/7405beaf-33c5-493b-bba7-83419a434632/volumes" Jan 06 15:04:45 crc kubenswrapper[4744]: I0106 15:04:45.328290 4744 generic.go:334] "Generic (PLEG): container finished" podID="aa8be235-0419-4ff3-8e76-d48ddbcfe47f" containerID="68c649576468951da258319abbce985baf1ed73a15d779d1e7b300baf1d3409a" exitCode=0 Jan 06 15:04:45 crc kubenswrapper[4744]: I0106 15:04:45.328407 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" event={"ID":"aa8be235-0419-4ff3-8e76-d48ddbcfe47f","Type":"ContainerDied","Data":"68c649576468951da258319abbce985baf1ed73a15d779d1e7b300baf1d3409a"} Jan 06 15:04:45 crc kubenswrapper[4744]: I0106 15:04:45.350986 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ptgpq" event={"ID":"0dcc5556-fb95-4735-b08c-a11e0e2d7042","Type":"ContainerStarted","Data":"5e1d7f929690066668c78fb2b00db034662be46a2b3f34992dcc7ff404ad3ec6"} Jan 06 15:04:45 crc kubenswrapper[4744]: I0106 15:04:45.407343 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-ptgpq" podStartSLOduration=1.762180108 podStartE2EDuration="48.407150602s" podCreationTimestamp="2026-01-06 15:03:57 +0000 UTC" firstStartedPulling="2026-01-06 15:03:58.025121734 +0000 UTC m=+1634.652588052" lastFinishedPulling="2026-01-06 15:04:44.670092228 +0000 UTC m=+1681.297558546" observedRunningTime="2026-01-06 15:04:45.389363651 +0000 UTC m=+1682.016829999" watchObservedRunningTime="2026-01-06 15:04:45.407150602 +0000 UTC m=+1682.034616920" Jan 06 15:04:45 crc kubenswrapper[4744]: I0106 15:04:45.810705 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 06 15:04:46 crc kubenswrapper[4744]: I0106 15:04:46.364750 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" event={"ID":"aa8be235-0419-4ff3-8e76-d48ddbcfe47f","Type":"ContainerStarted","Data":"715fc64dd2f86531c990681abf09a77873c94a07eb13c87e25fa5f529460e44c"} Jan 06 15:04:47 crc kubenswrapper[4744]: I0106 15:04:47.378475 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:47 crc kubenswrapper[4744]: I0106 15:04:47.415397 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" podStartSLOduration=9.415372912 podStartE2EDuration="9.415372912s" podCreationTimestamp="2026-01-06 15:04:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:04:47.403459407 +0000 UTC m=+1684.030925775" watchObservedRunningTime="2026-01-06 15:04:47.415372912 +0000 UTC m=+1684.042839230" Jan 06 15:04:47 crc kubenswrapper[4744]: I0106 15:04:47.712259 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:04:47 crc kubenswrapper[4744]: E0106 15:04:47.713115 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:04:50 crc kubenswrapper[4744]: I0106 15:04:50.426246 4744 generic.go:334] "Generic (PLEG): container finished" podID="0dcc5556-fb95-4735-b08c-a11e0e2d7042" containerID="5e1d7f929690066668c78fb2b00db034662be46a2b3f34992dcc7ff404ad3ec6" exitCode=0 Jan 06 15:04:50 crc kubenswrapper[4744]: I0106 15:04:50.426285 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ptgpq" event={"ID":"0dcc5556-fb95-4735-b08c-a11e0e2d7042","Type":"ContainerDied","Data":"5e1d7f929690066668c78fb2b00db034662be46a2b3f34992dcc7ff404ad3ec6"} Jan 06 15:04:50 crc kubenswrapper[4744]: I0106 15:04:50.432894 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"32ce34b0-736e-4d20-9c3b-4c76656092dc","Type":"ContainerStarted","Data":"eb192f9a3ac12a5515aa64bf916d1e3838572b1aab5a52769e54aa3646397448"} Jan 06 15:04:50 crc kubenswrapper[4744]: I0106 15:04:50.494649 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.883567191 podStartE2EDuration="46.494625043s" podCreationTimestamp="2026-01-06 15:04:04 +0000 UTC" firstStartedPulling="2026-01-06 15:04:05.590445214 +0000 UTC m=+1642.217911542" lastFinishedPulling="2026-01-06 15:04:49.201503036 +0000 UTC m=+1685.828969394" observedRunningTime="2026-01-06 15:04:50.481492566 +0000 UTC m=+1687.108958884" watchObservedRunningTime="2026-01-06 15:04:50.494625043 +0000 UTC m=+1687.122091391" Jan 06 15:04:51 crc kubenswrapper[4744]: I0106 15:04:51.913552 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ptgpq" Jan 06 15:04:51 crc kubenswrapper[4744]: I0106 15:04:51.966973 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-config-data\") pod \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " Jan 06 15:04:51 crc kubenswrapper[4744]: I0106 15:04:51.967097 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfsn4\" (UniqueName: \"kubernetes.io/projected/0dcc5556-fb95-4735-b08c-a11e0e2d7042-kube-api-access-vfsn4\") pod \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " Jan 06 15:04:51 crc kubenswrapper[4744]: I0106 15:04:51.967154 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-combined-ca-bundle\") pod \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\" (UID: \"0dcc5556-fb95-4735-b08c-a11e0e2d7042\") " Jan 06 15:04:51 crc kubenswrapper[4744]: I0106 15:04:51.974881 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dcc5556-fb95-4735-b08c-a11e0e2d7042-kube-api-access-vfsn4" (OuterVolumeSpecName: "kube-api-access-vfsn4") pod "0dcc5556-fb95-4735-b08c-a11e0e2d7042" (UID: "0dcc5556-fb95-4735-b08c-a11e0e2d7042"). InnerVolumeSpecName "kube-api-access-vfsn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:52 crc kubenswrapper[4744]: I0106 15:04:52.018359 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0dcc5556-fb95-4735-b08c-a11e0e2d7042" (UID: "0dcc5556-fb95-4735-b08c-a11e0e2d7042"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:04:52 crc kubenswrapper[4744]: I0106 15:04:52.060496 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-config-data" (OuterVolumeSpecName: "config-data") pod "0dcc5556-fb95-4735-b08c-a11e0e2d7042" (UID: "0dcc5556-fb95-4735-b08c-a11e0e2d7042"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:04:52 crc kubenswrapper[4744]: I0106 15:04:52.070960 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:52 crc kubenswrapper[4744]: I0106 15:04:52.071024 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfsn4\" (UniqueName: \"kubernetes.io/projected/0dcc5556-fb95-4735-b08c-a11e0e2d7042-kube-api-access-vfsn4\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:52 crc kubenswrapper[4744]: I0106 15:04:52.071036 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0dcc5556-fb95-4735-b08c-a11e0e2d7042-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:52 crc kubenswrapper[4744]: I0106 15:04:52.455258 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-ptgpq" event={"ID":"0dcc5556-fb95-4735-b08c-a11e0e2d7042","Type":"ContainerDied","Data":"d3f644f748e166b299521be819aaa2ae9b1d3176fe513c495cf5dbbe78ddfffc"} Jan 06 15:04:52 crc kubenswrapper[4744]: I0106 15:04:52.455299 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3f644f748e166b299521be819aaa2ae9b1d3176fe513c495cf5dbbe78ddfffc" Jan 06 15:04:52 crc kubenswrapper[4744]: I0106 15:04:52.455348 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-ptgpq" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.420306 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-5c5c49fd76-ww78j"] Jan 06 15:04:53 crc kubenswrapper[4744]: E0106 15:04:53.421278 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7405beaf-33c5-493b-bba7-83419a434632" containerName="init" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.421314 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7405beaf-33c5-493b-bba7-83419a434632" containerName="init" Jan 06 15:04:53 crc kubenswrapper[4744]: E0106 15:04:53.421352 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7405beaf-33c5-493b-bba7-83419a434632" containerName="dnsmasq-dns" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.421361 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7405beaf-33c5-493b-bba7-83419a434632" containerName="dnsmasq-dns" Jan 06 15:04:53 crc kubenswrapper[4744]: E0106 15:04:53.421391 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dcc5556-fb95-4735-b08c-a11e0e2d7042" containerName="heat-db-sync" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.421400 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dcc5556-fb95-4735-b08c-a11e0e2d7042" containerName="heat-db-sync" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.421671 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7405beaf-33c5-493b-bba7-83419a434632" containerName="dnsmasq-dns" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.421712 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dcc5556-fb95-4735-b08c-a11e0e2d7042" containerName="heat-db-sync" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.422837 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.466857 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5c5c49fd76-ww78j"] Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.498231 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-8fcb46fc5-bk6ql"] Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.500112 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.503373 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5sml\" (UniqueName: \"kubernetes.io/projected/39cc6313-b52f-4b01-b593-41008188e011-kube-api-access-v5sml\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.503476 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39cc6313-b52f-4b01-b593-41008188e011-config-data-custom\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.503499 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cc6313-b52f-4b01-b593-41008188e011-combined-ca-bundle\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.503531 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cc6313-b52f-4b01-b593-41008188e011-config-data\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.521345 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-db66db7b6-ghcg8"] Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.523024 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.579470 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-db66db7b6-ghcg8"] Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.606814 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-combined-ca-bundle\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.606906 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-public-tls-certs\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.606938 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdxsh\" (UniqueName: \"kubernetes.io/projected/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-kube-api-access-qdxsh\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.606978 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-combined-ca-bundle\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5sml\" (UniqueName: \"kubernetes.io/projected/39cc6313-b52f-4b01-b593-41008188e011-kube-api-access-v5sml\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607181 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-config-data\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607216 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-public-tls-certs\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607259 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxwpn\" (UniqueName: \"kubernetes.io/projected/fd669f35-602e-4680-a4ec-2d9bb7e66fca-kube-api-access-jxwpn\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607297 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39cc6313-b52f-4b01-b593-41008188e011-config-data-custom\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cc6313-b52f-4b01-b593-41008188e011-combined-ca-bundle\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607358 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-config-data\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607381 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-config-data-custom\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607405 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-config-data-custom\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607430 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cc6313-b52f-4b01-b593-41008188e011-config-data\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607482 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-internal-tls-certs\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.607515 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-internal-tls-certs\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.610270 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8fcb46fc5-bk6ql"] Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.614551 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39cc6313-b52f-4b01-b593-41008188e011-combined-ca-bundle\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.617001 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39cc6313-b52f-4b01-b593-41008188e011-config-data\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.627592 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/39cc6313-b52f-4b01-b593-41008188e011-config-data-custom\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.634349 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5sml\" (UniqueName: \"kubernetes.io/projected/39cc6313-b52f-4b01-b593-41008188e011-kube-api-access-v5sml\") pod \"heat-engine-5c5c49fd76-ww78j\" (UID: \"39cc6313-b52f-4b01-b593-41008188e011\") " pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.709897 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-public-tls-certs\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710139 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdxsh\" (UniqueName: \"kubernetes.io/projected/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-kube-api-access-qdxsh\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-combined-ca-bundle\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710295 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-config-data\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710317 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-public-tls-certs\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710348 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxwpn\" (UniqueName: \"kubernetes.io/projected/fd669f35-602e-4680-a4ec-2d9bb7e66fca-kube-api-access-jxwpn\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710379 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-config-data\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710395 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-config-data-custom\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710413 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-config-data-custom\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710456 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-internal-tls-certs\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710500 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-internal-tls-certs\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.710535 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-combined-ca-bundle\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.714838 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-combined-ca-bundle\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.719091 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-public-tls-certs\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.720032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-internal-tls-certs\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.720881 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-config-data-custom\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.721800 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-config-data\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.723008 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-internal-tls-certs\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.725664 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-combined-ca-bundle\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.726697 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-public-tls-certs\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.727143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-config-data\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.735039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxwpn\" (UniqueName: \"kubernetes.io/projected/fd669f35-602e-4680-a4ec-2d9bb7e66fca-kube-api-access-jxwpn\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.735358 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd669f35-602e-4680-a4ec-2d9bb7e66fca-config-data-custom\") pod \"heat-cfnapi-8fcb46fc5-bk6ql\" (UID: \"fd669f35-602e-4680-a4ec-2d9bb7e66fca\") " pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.737798 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdxsh\" (UniqueName: \"kubernetes.io/projected/0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0-kube-api-access-qdxsh\") pod \"heat-api-db66db7b6-ghcg8\" (UID: \"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0\") " pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.757048 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.821167 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:04:53 crc kubenswrapper[4744]: I0106 15:04:53.846233 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.174332 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5596c69fcc-nb6pm" Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.233985 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-tdwsz"] Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.234222 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" podUID="a931a95f-eaca-43d5-93ed-3d66ade2f347" containerName="dnsmasq-dns" containerID="cri-o://bd4bc99d5a0c08d8c164555b9afb38f0bb04c9142db681d0a143da501a3f7805" gracePeriod=10 Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.324694 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-5c5c49fd76-ww78j"] Jan 06 15:04:54 crc kubenswrapper[4744]: W0106 15:04:54.359251 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39cc6313_b52f_4b01_b593_41008188e011.slice/crio-0d36d6fddaa005fe04628646c402dae8fa8796684c915f632817c456cdcac7d7 WatchSource:0}: Error finding container 0d36d6fddaa005fe04628646c402dae8fa8796684c915f632817c456cdcac7d7: Status 404 returned error can't find the container with id 0d36d6fddaa005fe04628646c402dae8fa8796684c915f632817c456cdcac7d7 Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.490345 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-db66db7b6-ghcg8"] Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.496220 4744 generic.go:334] "Generic (PLEG): container finished" podID="a931a95f-eaca-43d5-93ed-3d66ade2f347" containerID="bd4bc99d5a0c08d8c164555b9afb38f0bb04c9142db681d0a143da501a3f7805" exitCode=0 Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.496287 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" event={"ID":"a931a95f-eaca-43d5-93ed-3d66ade2f347","Type":"ContainerDied","Data":"bd4bc99d5a0c08d8c164555b9afb38f0bb04c9142db681d0a143da501a3f7805"} Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.498331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5c5c49fd76-ww78j" event={"ID":"39cc6313-b52f-4b01-b593-41008188e011","Type":"ContainerStarted","Data":"0d36d6fddaa005fe04628646c402dae8fa8796684c915f632817c456cdcac7d7"} Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.501801 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8fcb46fc5-bk6ql"] Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.880770 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.946874 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgzgw\" (UniqueName: \"kubernetes.io/projected/a931a95f-eaca-43d5-93ed-3d66ade2f347-kube-api-access-dgzgw\") pod \"a931a95f-eaca-43d5-93ed-3d66ade2f347\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.946958 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-svc\") pod \"a931a95f-eaca-43d5-93ed-3d66ade2f347\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.947012 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-sb\") pod \"a931a95f-eaca-43d5-93ed-3d66ade2f347\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.947207 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-config\") pod \"a931a95f-eaca-43d5-93ed-3d66ade2f347\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.947240 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-openstack-edpm-ipam\") pod \"a931a95f-eaca-43d5-93ed-3d66ade2f347\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.947348 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-swift-storage-0\") pod \"a931a95f-eaca-43d5-93ed-3d66ade2f347\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.947485 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-nb\") pod \"a931a95f-eaca-43d5-93ed-3d66ade2f347\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " Jan 06 15:04:54 crc kubenswrapper[4744]: I0106 15:04:54.973667 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a931a95f-eaca-43d5-93ed-3d66ade2f347-kube-api-access-dgzgw" (OuterVolumeSpecName: "kube-api-access-dgzgw") pod "a931a95f-eaca-43d5-93ed-3d66ade2f347" (UID: "a931a95f-eaca-43d5-93ed-3d66ade2f347"). InnerVolumeSpecName "kube-api-access-dgzgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.030370 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a931a95f-eaca-43d5-93ed-3d66ade2f347" (UID: "a931a95f-eaca-43d5-93ed-3d66ade2f347"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.038062 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-config" (OuterVolumeSpecName: "config") pod "a931a95f-eaca-43d5-93ed-3d66ade2f347" (UID: "a931a95f-eaca-43d5-93ed-3d66ade2f347"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.048682 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a931a95f-eaca-43d5-93ed-3d66ade2f347" (UID: "a931a95f-eaca-43d5-93ed-3d66ade2f347"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.049284 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "a931a95f-eaca-43d5-93ed-3d66ade2f347" (UID: "a931a95f-eaca-43d5-93ed-3d66ade2f347"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.054066 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a931a95f-eaca-43d5-93ed-3d66ade2f347" (UID: "a931a95f-eaca-43d5-93ed-3d66ade2f347"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.056692 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-nb\") pod \"a931a95f-eaca-43d5-93ed-3d66ade2f347\" (UID: \"a931a95f-eaca-43d5-93ed-3d66ade2f347\") " Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.057641 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgzgw\" (UniqueName: \"kubernetes.io/projected/a931a95f-eaca-43d5-93ed-3d66ade2f347-kube-api-access-dgzgw\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.057662 4744 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.057672 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.057681 4744 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-config\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.057690 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:55 crc kubenswrapper[4744]: W0106 15:04:55.057753 4744 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/a931a95f-eaca-43d5-93ed-3d66ade2f347/volumes/kubernetes.io~configmap/ovsdbserver-nb Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.057763 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a931a95f-eaca-43d5-93ed-3d66ade2f347" (UID: "a931a95f-eaca-43d5-93ed-3d66ade2f347"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.074188 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a931a95f-eaca-43d5-93ed-3d66ade2f347" (UID: "a931a95f-eaca-43d5-93ed-3d66ade2f347"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.160730 4744 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.160766 4744 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a931a95f-eaca-43d5-93ed-3d66ade2f347-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.510848 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" event={"ID":"fd669f35-602e-4680-a4ec-2d9bb7e66fca","Type":"ContainerStarted","Data":"33cff2304295bd724ee563f253fba5c6d4e5618f320c228002a1a0c360ed116b"} Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.513832 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" event={"ID":"a931a95f-eaca-43d5-93ed-3d66ade2f347","Type":"ContainerDied","Data":"30044594a919137732b849710d82397327544e29babed63976fd961d083b65ea"} Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.513889 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-594cb89c79-tdwsz" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.513913 4744 scope.go:117] "RemoveContainer" containerID="bd4bc99d5a0c08d8c164555b9afb38f0bb04c9142db681d0a143da501a3f7805" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.518993 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-5c5c49fd76-ww78j" event={"ID":"39cc6313-b52f-4b01-b593-41008188e011","Type":"ContainerStarted","Data":"2c777b4b3e760cc44dc89bc545c3e0e121a2536e27bdcf771cbbd8c687e741a4"} Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.519366 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.525415 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-db66db7b6-ghcg8" event={"ID":"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0","Type":"ContainerStarted","Data":"eeb08467f2dc266c4b6149ca3ab5b0fc45908eba9476f3aee672ffbb5da7c94f"} Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.549618 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-5c5c49fd76-ww78j" podStartSLOduration=2.5495821149999998 podStartE2EDuration="2.549582115s" podCreationTimestamp="2026-01-06 15:04:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:04:55.53277507 +0000 UTC m=+1692.160241388" watchObservedRunningTime="2026-01-06 15:04:55.549582115 +0000 UTC m=+1692.177048433" Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.565262 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-tdwsz"] Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.577075 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-594cb89c79-tdwsz"] Jan 06 15:04:55 crc kubenswrapper[4744]: I0106 15:04:55.732004 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a931a95f-eaca-43d5-93ed-3d66ade2f347" path="/var/lib/kubelet/pods/a931a95f-eaca-43d5-93ed-3d66ade2f347/volumes" Jan 06 15:04:56 crc kubenswrapper[4744]: I0106 15:04:56.262811 4744 scope.go:117] "RemoveContainer" containerID="f9ff112feeafafd32e74ee1f6fee7dbe0710eff3ea7d9f186164339406308e1c" Jan 06 15:04:58 crc kubenswrapper[4744]: I0106 15:04:58.564083 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" event={"ID":"fd669f35-602e-4680-a4ec-2d9bb7e66fca","Type":"ContainerStarted","Data":"eee75f5aa7e6eb815c3c8854385698911313b8d66b8ce65b4b545c2510a83ad4"} Jan 06 15:04:58 crc kubenswrapper[4744]: I0106 15:04:58.567861 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-db66db7b6-ghcg8" event={"ID":"0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0","Type":"ContainerStarted","Data":"4e37912f14a9270aaf03aa29dffe3959e7111ae46442bab22f86c48b0d867a5d"} Jan 06 15:04:58 crc kubenswrapper[4744]: I0106 15:04:58.568715 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:04:58 crc kubenswrapper[4744]: I0106 15:04:58.592437 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" podStartSLOduration=3.632548692 podStartE2EDuration="5.592414173s" podCreationTimestamp="2026-01-06 15:04:53 +0000 UTC" firstStartedPulling="2026-01-06 15:04:54.504843 +0000 UTC m=+1691.132309318" lastFinishedPulling="2026-01-06 15:04:56.464708481 +0000 UTC m=+1693.092174799" observedRunningTime="2026-01-06 15:04:58.587317138 +0000 UTC m=+1695.214783466" watchObservedRunningTime="2026-01-06 15:04:58.592414173 +0000 UTC m=+1695.219880501" Jan 06 15:04:58 crc kubenswrapper[4744]: I0106 15:04:58.612304 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-db66db7b6-ghcg8" podStartSLOduration=3.646776178 podStartE2EDuration="5.612283019s" podCreationTimestamp="2026-01-06 15:04:53 +0000 UTC" firstStartedPulling="2026-01-06 15:04:54.498090491 +0000 UTC m=+1691.125556809" lastFinishedPulling="2026-01-06 15:04:56.463597332 +0000 UTC m=+1693.091063650" observedRunningTime="2026-01-06 15:04:58.602595222 +0000 UTC m=+1695.230061560" watchObservedRunningTime="2026-01-06 15:04:58.612283019 +0000 UTC m=+1695.239749337" Jan 06 15:04:58 crc kubenswrapper[4744]: I0106 15:04:58.825210 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:05:00 crc kubenswrapper[4744]: I0106 15:05:00.712721 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:05:00 crc kubenswrapper[4744]: E0106 15:05:00.713415 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:05:03 crc kubenswrapper[4744]: I0106 15:05:03.622687 4744 generic.go:334] "Generic (PLEG): container finished" podID="a33bb3f6-57ba-4e75-af0b-f9895afb79aa" containerID="ad125241967b0f11778000d67b4ebb9e7b4aadd6bfcd5632e40e41e1882dd4fd" exitCode=0 Jan 06 15:05:03 crc kubenswrapper[4744]: I0106 15:05:03.623097 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a33bb3f6-57ba-4e75-af0b-f9895afb79aa","Type":"ContainerDied","Data":"ad125241967b0f11778000d67b4ebb9e7b4aadd6bfcd5632e40e41e1882dd4fd"} Jan 06 15:05:05 crc kubenswrapper[4744]: I0106 15:05:05.203496 4744 scope.go:117] "RemoveContainer" containerID="614fe98476c53fb750be24913e7fb2bb300d66d103e2b2ab18c9ac222ce7b9e2" Jan 06 15:05:05 crc kubenswrapper[4744]: I0106 15:05:05.235659 4744 scope.go:117] "RemoveContainer" containerID="2f399288422bab7afd8aec2f805890b5db03b5fcf6c7ac7b4d96ca779c4de80a" Jan 06 15:05:05 crc kubenswrapper[4744]: I0106 15:05:05.649451 4744 generic.go:334] "Generic (PLEG): container finished" podID="b4c3d9bb-3170-4372-8cdd-ec1bfe092237" containerID="052672837a8e2c73135ae5ac6f5cb308335022fab3b7fc9052df9a1db50a9e61" exitCode=0 Jan 06 15:05:05 crc kubenswrapper[4744]: I0106 15:05:05.649517 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"b4c3d9bb-3170-4372-8cdd-ec1bfe092237","Type":"ContainerDied","Data":"052672837a8e2c73135ae5ac6f5cb308335022fab3b7fc9052df9a1db50a9e61"} Jan 06 15:05:06 crc kubenswrapper[4744]: I0106 15:05:06.515435 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-qzjdj" podUID="27d52d30-370e-44a9-82b1-eb0d7e197220" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:05:07 crc kubenswrapper[4744]: I0106 15:05:07.826712 4744 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.118768486s: [/var/lib/containers/storage/overlay/964ca9234e7b7e35e4824e033af1a708dc4f0152180abb53fe7eeffe554685a1/diff /var/log/pods/openshift-logging_logging-loki-ingester-0_631bc775-d5be-445e-abd6-d6ec8afd9cf2/loki-ingester/0.log]; will not log again for this container unless duration exceeds 2s Jan 06 15:05:07 crc kubenswrapper[4744]: I0106 15:05:07.835236 4744 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.127510588s: [/var/lib/containers/storage/overlay/8b5861e6371028462d3dffff2e5219f8fae084e34a27f015e57c47a8e822d117/diff /var/log/pods/openshift-logging_logging-loki-querier-76788598db-v9bbz_a2c75046-2bab-4e58-b54a-0172540755a2/loki-querier/0.log]; will not log again for this container unless duration exceeds 2s Jan 06 15:05:07 crc kubenswrapper[4744]: I0106 15:05:07.836595 4744 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.128905585s: [/var/lib/containers/storage/overlay/ff70d507803d1fe90cea625b98b7efe35ad0ab098be5c1c8bd42cb2060763126/diff /var/log/pods/openshift-logging_logging-loki-distributor-5f678c8dd6-jtrh9_7549f443-f7d9-42fe-94a0-5bc9565caa14/loki-distributor/0.log]; will not log again for this container unless duration exceeds 2s Jan 06 15:05:08 crc kubenswrapper[4744]: I0106 15:05:08.368948 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-db66db7b6-ghcg8" Jan 06 15:05:08 crc kubenswrapper[4744]: I0106 15:05:08.430573 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5985d564f7-8vfg7"] Jan 06 15:05:08 crc kubenswrapper[4744]: I0106 15:05:08.431069 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-5985d564f7-8vfg7" podUID="83192382-3f34-4716-b10b-123923752ba5" containerName="heat-api" containerID="cri-o://7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4" gracePeriod=60 Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.777890 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-8fcb46fc5-bk6ql" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.859917 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h"] Jan 06 15:05:09 crc kubenswrapper[4744]: E0106 15:05:09.860824 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a931a95f-eaca-43d5-93ed-3d66ade2f347" containerName="dnsmasq-dns" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.860930 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a931a95f-eaca-43d5-93ed-3d66ade2f347" containerName="dnsmasq-dns" Jan 06 15:05:09 crc kubenswrapper[4744]: E0106 15:05:09.861094 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a931a95f-eaca-43d5-93ed-3d66ade2f347" containerName="init" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.861185 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a931a95f-eaca-43d5-93ed-3d66ade2f347" containerName="init" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.861599 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a931a95f-eaca-43d5-93ed-3d66ade2f347" containerName="dnsmasq-dns" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.862686 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.867083 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.867312 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.867481 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.867989 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.880411 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7f5fdd5db8-d2k9t"] Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.880643 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" podUID="4adab37a-4969-424b-87dd-5e59dc9cd756" containerName="heat-cfnapi" containerID="cri-o://cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086" gracePeriod=60 Jan 06 15:05:09 crc kubenswrapper[4744]: I0106 15:05:09.936934 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h"] Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.005184 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.005253 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.005291 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6zzj\" (UniqueName: \"kubernetes.io/projected/42eceb98-5fea-4ae7-827f-94b867781272-kube-api-access-l6zzj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.005762 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.108129 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.108424 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.108517 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.108599 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6zzj\" (UniqueName: \"kubernetes.io/projected/42eceb98-5fea-4ae7-827f-94b867781272-kube-api-access-l6zzj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.125202 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.133075 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.137641 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.148366 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6zzj\" (UniqueName: \"kubernetes.io/projected/42eceb98-5fea-4ae7-827f-94b867781272-kube-api-access-l6zzj\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.185658 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.739592 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a33bb3f6-57ba-4e75-af0b-f9895afb79aa","Type":"ContainerStarted","Data":"664486a94239aade45080f12f3a7ec2b9c9ae40b87da4235032c8704b732d307"} Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.740010 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.743661 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"b4c3d9bb-3170-4372-8cdd-ec1bfe092237","Type":"ContainerStarted","Data":"b67010788d3556d7d601af20dcd145c81fbfdac34ebd9de1e37b2bd201806ec5"} Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.743926 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Jan 06 15:05:10 crc kubenswrapper[4744]: I0106 15:05:10.773585 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=56.773560344 podStartE2EDuration="56.773560344s" podCreationTimestamp="2026-01-06 15:04:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:05:10.766024045 +0000 UTC m=+1707.393490383" watchObservedRunningTime="2026-01-06 15:05:10.773560344 +0000 UTC m=+1707.401026672" Jan 06 15:05:11 crc kubenswrapper[4744]: I0106 15:05:11.279112 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=43.279093891 podStartE2EDuration="43.279093891s" podCreationTimestamp="2026-01-06 15:04:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:05:10.811217541 +0000 UTC m=+1707.438683859" watchObservedRunningTime="2026-01-06 15:05:11.279093891 +0000 UTC m=+1707.906560219" Jan 06 15:05:11 crc kubenswrapper[4744]: I0106 15:05:11.280336 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h"] Jan 06 15:05:11 crc kubenswrapper[4744]: W0106 15:05:11.283806 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42eceb98_5fea_4ae7_827f_94b867781272.slice/crio-29cd3f48ede87559d88f380dafecb3fd6526d376aa736161ef95c53ddc31b2b4 WatchSource:0}: Error finding container 29cd3f48ede87559d88f380dafecb3fd6526d376aa736161ef95c53ddc31b2b4: Status 404 returned error can't find the container with id 29cd3f48ede87559d88f380dafecb3fd6526d376aa736161ef95c53ddc31b2b4 Jan 06 15:05:11 crc kubenswrapper[4744]: I0106 15:05:11.772595 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" event={"ID":"42eceb98-5fea-4ae7-827f-94b867781272","Type":"ContainerStarted","Data":"29cd3f48ede87559d88f380dafecb3fd6526d376aa736161ef95c53ddc31b2b4"} Jan 06 15:05:12 crc kubenswrapper[4744]: I0106 15:05:12.871215 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5985d564f7-8vfg7" podUID="83192382-3f34-4716-b10b-123923752ba5" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.228:8004/healthcheck\": read tcp 10.217.0.2:59392->10.217.0.228:8004: read: connection reset by peer" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.338316 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" podUID="4adab37a-4969-424b-87dd-5e59dc9cd756" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.229:8000/healthcheck\": read tcp 10.217.0.2:32876->10.217.0.229:8000: read: connection reset by peer" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.591767 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.745857 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-internal-tls-certs\") pod \"83192382-3f34-4716-b10b-123923752ba5\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.746264 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data\") pod \"83192382-3f34-4716-b10b-123923752ba5\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.746425 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjd2g\" (UniqueName: \"kubernetes.io/projected/83192382-3f34-4716-b10b-123923752ba5-kube-api-access-pjd2g\") pod \"83192382-3f34-4716-b10b-123923752ba5\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.746519 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-public-tls-certs\") pod \"83192382-3f34-4716-b10b-123923752ba5\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.746734 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data-custom\") pod \"83192382-3f34-4716-b10b-123923752ba5\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.746771 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-combined-ca-bundle\") pod \"83192382-3f34-4716-b10b-123923752ba5\" (UID: \"83192382-3f34-4716-b10b-123923752ba5\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.768920 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83192382-3f34-4716-b10b-123923752ba5-kube-api-access-pjd2g" (OuterVolumeSpecName: "kube-api-access-pjd2g") pod "83192382-3f34-4716-b10b-123923752ba5" (UID: "83192382-3f34-4716-b10b-123923752ba5"). InnerVolumeSpecName "kube-api-access-pjd2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.770720 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "83192382-3f34-4716-b10b-123923752ba5" (UID: "83192382-3f34-4716-b10b-123923752ba5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.778367 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.796818 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83192382-3f34-4716-b10b-123923752ba5" (UID: "83192382-3f34-4716-b10b-123923752ba5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.803472 4744 generic.go:334] "Generic (PLEG): container finished" podID="83192382-3f34-4716-b10b-123923752ba5" containerID="7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4" exitCode=0 Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.803632 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5985d564f7-8vfg7" event={"ID":"83192382-3f34-4716-b10b-123923752ba5","Type":"ContainerDied","Data":"7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4"} Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.803659 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5985d564f7-8vfg7" event={"ID":"83192382-3f34-4716-b10b-123923752ba5","Type":"ContainerDied","Data":"c74309585cbe0903bbf74b22adecd118b02581ea824df10f453720a4601bfb90"} Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.803693 4744 scope.go:117] "RemoveContainer" containerID="7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.803818 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5985d564f7-8vfg7" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.809642 4744 generic.go:334] "Generic (PLEG): container finished" podID="4adab37a-4969-424b-87dd-5e59dc9cd756" containerID="cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086" exitCode=0 Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.809694 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" event={"ID":"4adab37a-4969-424b-87dd-5e59dc9cd756","Type":"ContainerDied","Data":"cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086"} Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.809722 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" event={"ID":"4adab37a-4969-424b-87dd-5e59dc9cd756","Type":"ContainerDied","Data":"caccc07ce56e2fcd8e093b6cf5655bfee83402876f1534a810c1d548a1f6f1e0"} Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.809788 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7f5fdd5db8-d2k9t" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.833508 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-5c5c49fd76-ww78j" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.847291 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "83192382-3f34-4716-b10b-123923752ba5" (UID: "83192382-3f34-4716-b10b-123923752ba5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.851291 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-combined-ca-bundle\") pod \"4adab37a-4969-424b-87dd-5e59dc9cd756\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.851514 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgwqr\" (UniqueName: \"kubernetes.io/projected/4adab37a-4969-424b-87dd-5e59dc9cd756-kube-api-access-xgwqr\") pod \"4adab37a-4969-424b-87dd-5e59dc9cd756\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.851795 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data-custom\") pod \"4adab37a-4969-424b-87dd-5e59dc9cd756\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.851912 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-public-tls-certs\") pod \"4adab37a-4969-424b-87dd-5e59dc9cd756\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.851994 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data\") pod \"4adab37a-4969-424b-87dd-5e59dc9cd756\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.852173 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-internal-tls-certs\") pod \"4adab37a-4969-424b-87dd-5e59dc9cd756\" (UID: \"4adab37a-4969-424b-87dd-5e59dc9cd756\") " Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.853016 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.853471 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjd2g\" (UniqueName: \"kubernetes.io/projected/83192382-3f34-4716-b10b-123923752ba5-kube-api-access-pjd2g\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.853545 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.853679 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.866370 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4adab37a-4969-424b-87dd-5e59dc9cd756-kube-api-access-xgwqr" (OuterVolumeSpecName: "kube-api-access-xgwqr") pod "4adab37a-4969-424b-87dd-5e59dc9cd756" (UID: "4adab37a-4969-424b-87dd-5e59dc9cd756"). InnerVolumeSpecName "kube-api-access-xgwqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.873293 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4adab37a-4969-424b-87dd-5e59dc9cd756" (UID: "4adab37a-4969-424b-87dd-5e59dc9cd756"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.905672 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-56d794f5-7hvdj"] Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.905880 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-56d794f5-7hvdj" podUID="f7bf873c-7171-4ec5-a141-aeabfb5501df" containerName="heat-engine" containerID="cri-o://d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf" gracePeriod=60 Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.913066 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data" (OuterVolumeSpecName: "config-data") pod "83192382-3f34-4716-b10b-123923752ba5" (UID: "83192382-3f34-4716-b10b-123923752ba5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.922288 4744 scope.go:117] "RemoveContainer" containerID="7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4" Jan 06 15:05:13 crc kubenswrapper[4744]: E0106 15:05:13.927721 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4\": container with ID starting with 7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4 not found: ID does not exist" containerID="7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.927947 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4"} err="failed to get container status \"7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4\": rpc error: code = NotFound desc = could not find container \"7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4\": container with ID starting with 7482470fbf581811b0852722fe4cba17baf94c7d03fda36bca5e150094917fd4 not found: ID does not exist" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.928027 4744 scope.go:117] "RemoveContainer" containerID="cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.930427 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data" (OuterVolumeSpecName: "config-data") pod "4adab37a-4969-424b-87dd-5e59dc9cd756" (UID: "4adab37a-4969-424b-87dd-5e59dc9cd756"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.936197 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "83192382-3f34-4716-b10b-123923752ba5" (UID: "83192382-3f34-4716-b10b-123923752ba5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.950344 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4adab37a-4969-424b-87dd-5e59dc9cd756" (UID: "4adab37a-4969-424b-87dd-5e59dc9cd756"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.958579 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.958621 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgwqr\" (UniqueName: \"kubernetes.io/projected/4adab37a-4969-424b-87dd-5e59dc9cd756-kube-api-access-xgwqr\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.958636 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.958648 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.958659 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/83192382-3f34-4716-b10b-123923752ba5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.958669 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.987388 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4adab37a-4969-424b-87dd-5e59dc9cd756" (UID: "4adab37a-4969-424b-87dd-5e59dc9cd756"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:13 crc kubenswrapper[4744]: I0106 15:05:13.998811 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4adab37a-4969-424b-87dd-5e59dc9cd756" (UID: "4adab37a-4969-424b-87dd-5e59dc9cd756"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:14 crc kubenswrapper[4744]: I0106 15:05:14.030905 4744 scope.go:117] "RemoveContainer" containerID="cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086" Jan 06 15:05:14 crc kubenswrapper[4744]: E0106 15:05:14.031460 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086\": container with ID starting with cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086 not found: ID does not exist" containerID="cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086" Jan 06 15:05:14 crc kubenswrapper[4744]: I0106 15:05:14.031503 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086"} err="failed to get container status \"cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086\": rpc error: code = NotFound desc = could not find container \"cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086\": container with ID starting with cd0d6ca45776bd9487563ae5ab825b4f912973e70610924bc032290c306cf086 not found: ID does not exist" Jan 06 15:05:14 crc kubenswrapper[4744]: I0106 15:05:14.061929 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:14 crc kubenswrapper[4744]: I0106 15:05:14.061964 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4adab37a-4969-424b-87dd-5e59dc9cd756-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:14 crc kubenswrapper[4744]: I0106 15:05:14.148379 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5985d564f7-8vfg7"] Jan 06 15:05:14 crc kubenswrapper[4744]: I0106 15:05:14.202259 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5985d564f7-8vfg7"] Jan 06 15:05:14 crc kubenswrapper[4744]: I0106 15:05:14.220282 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7f5fdd5db8-d2k9t"] Jan 06 15:05:14 crc kubenswrapper[4744]: I0106 15:05:14.235913 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7f5fdd5db8-d2k9t"] Jan 06 15:05:14 crc kubenswrapper[4744]: I0106 15:05:14.711189 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:05:14 crc kubenswrapper[4744]: E0106 15:05:14.711529 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:05:15 crc kubenswrapper[4744]: I0106 15:05:15.728861 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4adab37a-4969-424b-87dd-5e59dc9cd756" path="/var/lib/kubelet/pods/4adab37a-4969-424b-87dd-5e59dc9cd756/volumes" Jan 06 15:05:15 crc kubenswrapper[4744]: I0106 15:05:15.729728 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83192382-3f34-4716-b10b-123923752ba5" path="/var/lib/kubelet/pods/83192382-3f34-4716-b10b-123923752ba5/volumes" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.271735 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-2mm6l"] Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.287482 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-2mm6l"] Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.331779 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-jndqv"] Jan 06 15:05:17 crc kubenswrapper[4744]: E0106 15:05:17.332283 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4adab37a-4969-424b-87dd-5e59dc9cd756" containerName="heat-cfnapi" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.332303 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4adab37a-4969-424b-87dd-5e59dc9cd756" containerName="heat-cfnapi" Jan 06 15:05:17 crc kubenswrapper[4744]: E0106 15:05:17.332319 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83192382-3f34-4716-b10b-123923752ba5" containerName="heat-api" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.332326 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="83192382-3f34-4716-b10b-123923752ba5" containerName="heat-api" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.332549 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4adab37a-4969-424b-87dd-5e59dc9cd756" containerName="heat-cfnapi" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.332569 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="83192382-3f34-4716-b10b-123923752ba5" containerName="heat-api" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.333319 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.340359 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.343041 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jndqv"] Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.459659 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fctmc\" (UniqueName: \"kubernetes.io/projected/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-kube-api-access-fctmc\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.459761 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-scripts\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.459859 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-combined-ca-bundle\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.460006 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-config-data\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.562784 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-config-data\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.563000 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fctmc\" (UniqueName: \"kubernetes.io/projected/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-kube-api-access-fctmc\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.563650 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-scripts\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.564319 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-combined-ca-bundle\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.568322 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-scripts\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.568609 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-config-data\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.575701 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-combined-ca-bundle\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.582212 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fctmc\" (UniqueName: \"kubernetes.io/projected/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-kube-api-access-fctmc\") pod \"aodh-db-sync-jndqv\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.654948 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jndqv" Jan 06 15:05:17 crc kubenswrapper[4744]: I0106 15:05:17.728371 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3361c675-133b-4206-86f1-879f52cdb5bb" path="/var/lib/kubelet/pods/3361c675-133b-4206-86f1-879f52cdb5bb/volumes" Jan 06 15:05:19 crc kubenswrapper[4744]: E0106 15:05:19.228877 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:05:19 crc kubenswrapper[4744]: E0106 15:05:19.230404 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:05:19 crc kubenswrapper[4744]: E0106 15:05:19.232205 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:05:19 crc kubenswrapper[4744]: E0106 15:05:19.232270 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-56d794f5-7hvdj" podUID="f7bf873c-7171-4ec5-a141-aeabfb5501df" containerName="heat-engine" Jan 06 15:05:24 crc kubenswrapper[4744]: I0106 15:05:24.165606 4744 patch_prober.go:28] interesting pod/logging-loki-gateway-b67674996-w4jmm container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 15:05:24 crc kubenswrapper[4744]: I0106 15:05:24.166441 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-b67674996-w4jmm" podUID="4e939cdc-a027-4c89-b0e8-f5c0dff2af63" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.56:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 06 15:05:24 crc kubenswrapper[4744]: I0106 15:05:24.930436 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-mmvx7" podUID="4d920867-2ba6-42f6-acc3-266cbfbc7716" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:05:25 crc kubenswrapper[4744]: I0106 15:05:25.757129 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="73416f0d-1377-4ac4-ab12-c5d54151325b" containerName="ovn-northd" probeResult="failure" output="command timed out" Jan 06 15:05:25 crc kubenswrapper[4744]: I0106 15:05:25.757358 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="73416f0d-1377-4ac4-ab12-c5d54151325b" containerName="ovn-northd" probeResult="failure" output="command timed out" Jan 06 15:05:26 crc kubenswrapper[4744]: I0106 15:05:26.332564 4744 patch_prober.go:28] interesting pod/route-controller-manager-64bb695654-t9m84 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 15:05:26 crc kubenswrapper[4744]: I0106 15:05:26.333777 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" podUID="8ed4bc19-0cbb-41fe-9956-5aef0d4c287b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 06 15:05:26 crc kubenswrapper[4744]: I0106 15:05:26.334127 4744 patch_prober.go:28] interesting pod/route-controller-manager-64bb695654-t9m84 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 15:05:26 crc kubenswrapper[4744]: I0106 15:05:26.334801 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-64bb695654-t9m84" podUID="8ed4bc19-0cbb-41fe-9956-5aef0d4c287b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 06 15:05:26 crc kubenswrapper[4744]: I0106 15:05:26.520765 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="a33bb3f6-57ba-4e75-af0b-f9895afb79aa" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.14:5671: connect: connection refused" Jan 06 15:05:26 crc kubenswrapper[4744]: I0106 15:05:26.694455 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-7b88bfc995-k8n6t" podUID="0abaa053-82ac-4b9f-a4d2-2bc5024a97eb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:05:27 crc kubenswrapper[4744]: I0106 15:05:27.301385 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-6c866cfdcb-8q8ks" podUID="65ec28e9-38c5-4972-b8b8-1447de87ee85" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:05:27 crc kubenswrapper[4744]: I0106 15:05:27.373113 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-7hrfh" podUID="458bdba1-55cb-49ad-81ef-5e4c8b07c920" containerName="registry-server" probeResult="failure" output=< Jan 06 15:05:27 crc kubenswrapper[4744]: timeout: health rpc did not complete within 1s Jan 06 15:05:27 crc kubenswrapper[4744]: > Jan 06 15:05:28 crc kubenswrapper[4744]: I0106 15:05:28.782642 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="b4c3d9bb-3170-4372-8cdd-ec1bfe092237" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.16:5671: connect: connection refused" Jan 06 15:05:28 crc kubenswrapper[4744]: E0106 15:05:28.903718 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest" Jan 06 15:05:28 crc kubenswrapper[4744]: E0106 15:05:28.903910 4744 kuberuntime_manager.go:1274] "Unhandled Error" err=< Jan 06 15:05:28 crc kubenswrapper[4744]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Jan 06 15:05:28 crc kubenswrapper[4744]: - hosts: all Jan 06 15:05:28 crc kubenswrapper[4744]: strategy: linear Jan 06 15:05:28 crc kubenswrapper[4744]: tasks: Jan 06 15:05:28 crc kubenswrapper[4744]: - name: Enable podified-repos Jan 06 15:05:28 crc kubenswrapper[4744]: become: true Jan 06 15:05:28 crc kubenswrapper[4744]: ansible.builtin.shell: | Jan 06 15:05:28 crc kubenswrapper[4744]: set -euxo pipefail Jan 06 15:05:28 crc kubenswrapper[4744]: pushd /var/tmp Jan 06 15:05:28 crc kubenswrapper[4744]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Jan 06 15:05:28 crc kubenswrapper[4744]: pushd repo-setup-main Jan 06 15:05:28 crc kubenswrapper[4744]: python3 -m venv ./venv Jan 06 15:05:28 crc kubenswrapper[4744]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Jan 06 15:05:28 crc kubenswrapper[4744]: ./venv/bin/repo-setup current-podified -b antelope Jan 06 15:05:28 crc kubenswrapper[4744]: popd Jan 06 15:05:28 crc kubenswrapper[4744]: rm -rf repo-setup-main Jan 06 15:05:28 crc kubenswrapper[4744]: Jan 06 15:05:28 crc kubenswrapper[4744]: Jan 06 15:05:28 crc kubenswrapper[4744]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Jan 06 15:05:28 crc kubenswrapper[4744]: edpm_override_hosts: openstack-edpm-ipam Jan 06 15:05:28 crc kubenswrapper[4744]: edpm_service_type: repo-setup Jan 06 15:05:28 crc kubenswrapper[4744]: Jan 06 15:05:28 crc kubenswrapper[4744]: Jan 06 15:05:28 crc kubenswrapper[4744]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key-openstack-edpm-ipam,ReadOnly:false,MountPath:/runner/env/ssh_key/ssh_key_openstack-edpm-ipam,SubPath:ssh_key_openstack-edpm-ipam,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l6zzj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h_openstack(42eceb98-5fea-4ae7-827f-94b867781272): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Jan 06 15:05:28 crc kubenswrapper[4744]: > logger="UnhandledError" Jan 06 15:05:28 crc kubenswrapper[4744]: E0106 15:05:28.905502 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" podUID="42eceb98-5fea-4ae7-827f-94b867781272" Jan 06 15:05:29 crc kubenswrapper[4744]: E0106 15:05:29.001737 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" podUID="42eceb98-5fea-4ae7-827f-94b867781272" Jan 06 15:05:29 crc kubenswrapper[4744]: E0106 15:05:29.230604 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:05:29 crc kubenswrapper[4744]: E0106 15:05:29.233070 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:05:29 crc kubenswrapper[4744]: E0106 15:05:29.234980 4744 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Jan 06 15:05:29 crc kubenswrapper[4744]: E0106 15:05:29.235068 4744 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-56d794f5-7hvdj" podUID="f7bf873c-7171-4ec5-a141-aeabfb5501df" containerName="heat-engine" Jan 06 15:05:29 crc kubenswrapper[4744]: I0106 15:05:29.712043 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:05:29 crc kubenswrapper[4744]: E0106 15:05:29.712864 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:05:30 crc kubenswrapper[4744]: I0106 15:05:30.754327 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="b2b4c15b-189d-41b2-b121-fdc25beb18a4" containerName="galera" probeResult="failure" output="command timed out" Jan 06 15:05:30 crc kubenswrapper[4744]: I0106 15:05:30.755862 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b2b4c15b-189d-41b2-b121-fdc25beb18a4" containerName="galera" probeResult="failure" output="command timed out" Jan 06 15:05:31 crc kubenswrapper[4744]: I0106 15:05:31.918450 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-jndqv"] Jan 06 15:05:32 crc kubenswrapper[4744]: I0106 15:05:32.044055 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jndqv" event={"ID":"df07cf75-bdf5-43e2-8066-e1b4f2d7b583","Type":"ContainerStarted","Data":"3d9cf0650e7d0fe5576b66380eca3260d86fd574a84ee2941b4ffb3c55920887"} Jan 06 15:05:35 crc kubenswrapper[4744]: I0106 15:05:35.086871 4744 generic.go:334] "Generic (PLEG): container finished" podID="f7bf873c-7171-4ec5-a141-aeabfb5501df" containerID="d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf" exitCode=0 Jan 06 15:05:35 crc kubenswrapper[4744]: I0106 15:05:35.087023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-56d794f5-7hvdj" event={"ID":"f7bf873c-7171-4ec5-a141-aeabfb5501df","Type":"ContainerDied","Data":"d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf"} Jan 06 15:05:36 crc kubenswrapper[4744]: I0106 15:05:36.522364 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 06 15:05:37 crc kubenswrapper[4744]: I0106 15:05:37.769113 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:05:37 crc kubenswrapper[4744]: I0106 15:05:37.898761 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data\") pod \"f7bf873c-7171-4ec5-a141-aeabfb5501df\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " Jan 06 15:05:37 crc kubenswrapper[4744]: I0106 15:05:37.899431 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data-custom\") pod \"f7bf873c-7171-4ec5-a141-aeabfb5501df\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " Jan 06 15:05:37 crc kubenswrapper[4744]: I0106 15:05:37.899725 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g2pd\" (UniqueName: \"kubernetes.io/projected/f7bf873c-7171-4ec5-a141-aeabfb5501df-kube-api-access-4g2pd\") pod \"f7bf873c-7171-4ec5-a141-aeabfb5501df\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " Jan 06 15:05:37 crc kubenswrapper[4744]: I0106 15:05:37.899821 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-combined-ca-bundle\") pod \"f7bf873c-7171-4ec5-a141-aeabfb5501df\" (UID: \"f7bf873c-7171-4ec5-a141-aeabfb5501df\") " Jan 06 15:05:37 crc kubenswrapper[4744]: I0106 15:05:37.907376 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f7bf873c-7171-4ec5-a141-aeabfb5501df" (UID: "f7bf873c-7171-4ec5-a141-aeabfb5501df"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:37 crc kubenswrapper[4744]: I0106 15:05:37.907585 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7bf873c-7171-4ec5-a141-aeabfb5501df-kube-api-access-4g2pd" (OuterVolumeSpecName: "kube-api-access-4g2pd") pod "f7bf873c-7171-4ec5-a141-aeabfb5501df" (UID: "f7bf873c-7171-4ec5-a141-aeabfb5501df"). InnerVolumeSpecName "kube-api-access-4g2pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:05:37 crc kubenswrapper[4744]: I0106 15:05:37.945110 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7bf873c-7171-4ec5-a141-aeabfb5501df" (UID: "f7bf873c-7171-4ec5-a141-aeabfb5501df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:37 crc kubenswrapper[4744]: I0106 15:05:37.963838 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data" (OuterVolumeSpecName: "config-data") pod "f7bf873c-7171-4ec5-a141-aeabfb5501df" (UID: "f7bf873c-7171-4ec5-a141-aeabfb5501df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.002913 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.002940 4744 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.002951 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g2pd\" (UniqueName: \"kubernetes.io/projected/f7bf873c-7171-4ec5-a141-aeabfb5501df-kube-api-access-4g2pd\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.002961 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7bf873c-7171-4ec5-a141-aeabfb5501df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.127022 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-56d794f5-7hvdj" event={"ID":"f7bf873c-7171-4ec5-a141-aeabfb5501df","Type":"ContainerDied","Data":"2d592034d00e95bc1ab7277831a57867116d1ff7bf6d8587949585a287b469e5"} Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.127376 4744 scope.go:117] "RemoveContainer" containerID="d1d0739a01d9664aa920397a7f0f2b148d4d8fd302fec33d8e6390f3c2857faf" Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.127133 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-56d794f5-7hvdj" Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.176923 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-56d794f5-7hvdj"] Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.193707 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-56d794f5-7hvdj"] Jan 06 15:05:38 crc kubenswrapper[4744]: I0106 15:05:38.777153 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="b4c3d9bb-3170-4372-8cdd-ec1bfe092237" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.16:5671: connect: connection refused" Jan 06 15:05:39 crc kubenswrapper[4744]: I0106 15:05:39.732795 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7bf873c-7171-4ec5-a141-aeabfb5501df" path="/var/lib/kubelet/pods/f7bf873c-7171-4ec5-a141-aeabfb5501df/volumes" Jan 06 15:05:41 crc kubenswrapper[4744]: I0106 15:05:41.711202 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:05:41 crc kubenswrapper[4744]: E0106 15:05:41.711975 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:05:48 crc kubenswrapper[4744]: I0106 15:05:48.778785 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Jan 06 15:05:48 crc kubenswrapper[4744]: I0106 15:05:48.845566 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Jan 06 15:05:49 crc kubenswrapper[4744]: E0106 15:05:49.380867 4744 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-aodh-api:current-tested" Jan 06 15:05:49 crc kubenswrapper[4744]: E0106 15:05:49.381255 4744 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-aodh-api:current-tested" Jan 06 15:05:49 crc kubenswrapper[4744]: E0106 15:05:49.381399 4744 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:aodh-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-aodh-api:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:AodhPassword,Value:,ValueFrom:&EnvVarSource{FieldRef:nil,ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:&SecretKeySelector{LocalObjectReference:LocalObjectReference{Name:osp-secret,},Key:AodhPassword,Optional:nil,},},},EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:aodh-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fctmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42402,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod aodh-db-sync-jndqv_openstack(df07cf75-bdf5-43e2-8066-e1b4f2d7b583): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 06 15:05:49 crc kubenswrapper[4744]: E0106 15:05:49.382694 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"aodh-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/aodh-db-sync-jndqv" podUID="df07cf75-bdf5-43e2-8066-e1b4f2d7b583" Jan 06 15:05:49 crc kubenswrapper[4744]: I0106 15:05:49.550672 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:05:50 crc kubenswrapper[4744]: I0106 15:05:50.288288 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" event={"ID":"42eceb98-5fea-4ae7-827f-94b867781272","Type":"ContainerStarted","Data":"76d97059b01ffa66b999ea095cb76b4a2642e0a63f949bff76b983c119382f09"} Jan 06 15:05:50 crc kubenswrapper[4744]: E0106 15:05:50.291633 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"aodh-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-aodh-api:current-tested\\\"\"" pod="openstack/aodh-db-sync-jndqv" podUID="df07cf75-bdf5-43e2-8066-e1b4f2d7b583" Jan 06 15:05:50 crc kubenswrapper[4744]: I0106 15:05:50.313669 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" podStartSLOduration=3.052746584 podStartE2EDuration="41.313651883s" podCreationTimestamp="2026-01-06 15:05:09 +0000 UTC" firstStartedPulling="2026-01-06 15:05:11.287273658 +0000 UTC m=+1707.914739976" lastFinishedPulling="2026-01-06 15:05:49.548178957 +0000 UTC m=+1746.175645275" observedRunningTime="2026-01-06 15:05:50.302802796 +0000 UTC m=+1746.930269114" watchObservedRunningTime="2026-01-06 15:05:50.313651883 +0000 UTC m=+1746.941118221" Jan 06 15:05:51 crc kubenswrapper[4744]: I0106 15:05:51.985252 4744 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.02831258s: [/var/lib/containers/storage/overlay/44f7eadd381f170852997be2297a02374c60203e4c760a0828a9722f11499f50/diff /var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_e85ae009-0426-4c0e-a1af-de7cd66148cb/prometheus-operator-admission-webhook/0.log]; will not log again for this container unless duration exceeds 2s Jan 06 15:05:53 crc kubenswrapper[4744]: I0106 15:05:53.538133 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-1" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="rabbitmq" containerID="cri-o://98ff97c26acef954ef996e4aacd4e7dec375b11833016c6e200cd86dd7f206d4" gracePeriod=604796 Jan 06 15:05:54 crc kubenswrapper[4744]: I0106 15:05:54.711851 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:05:54 crc kubenswrapper[4744]: E0106 15:05:54.712451 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:05:58 crc kubenswrapper[4744]: I0106 15:05:58.729309 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.420720 4744 generic.go:334] "Generic (PLEG): container finished" podID="bef00f62-0142-4a72-9849-12246c13d004" containerID="98ff97c26acef954ef996e4aacd4e7dec375b11833016c6e200cd86dd7f206d4" exitCode=0 Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.420818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"bef00f62-0142-4a72-9849-12246c13d004","Type":"ContainerDied","Data":"98ff97c26acef954ef996e4aacd4e7dec375b11833016c6e200cd86dd7f206d4"} Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.421408 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"bef00f62-0142-4a72-9849-12246c13d004","Type":"ContainerDied","Data":"183a0a8f321be07ca3156bc0309d0d413397f55c64de37ac737fa0da11c0ac2b"} Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.421430 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="183a0a8f321be07ca3156bc0309d0d413397f55c64de37ac737fa0da11c0ac2b" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.509630 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625036 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hhmz\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-kube-api-access-7hhmz\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625150 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bef00f62-0142-4a72-9849-12246c13d004-pod-info\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625227 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-plugins\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625255 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-confd\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625594 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625702 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-server-conf\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625797 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-tls\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625858 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-config-data\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625882 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bef00f62-0142-4a72-9849-12246c13d004-erlang-cookie-secret\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625918 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-erlang-cookie\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.625972 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-plugins-conf\") pod \"bef00f62-0142-4a72-9849-12246c13d004\" (UID: \"bef00f62-0142-4a72-9849-12246c13d004\") " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.626117 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.628542 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.628691 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.630451 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.633538 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/bef00f62-0142-4a72-9849-12246c13d004-pod-info" (OuterVolumeSpecName: "pod-info") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.635608 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef00f62-0142-4a72-9849-12246c13d004-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.635662 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-kube-api-access-7hhmz" (OuterVolumeSpecName: "kube-api-access-7hhmz") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "kube-api-access-7hhmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.638515 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.668930 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6" (OuterVolumeSpecName: "persistence") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "pvc-16439a32-bb72-4737-af92-d59e2f11ecb6". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.672323 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-config-data" (OuterVolumeSpecName: "config-data") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.746741 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-server-conf" (OuterVolumeSpecName: "server-conf") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.756226 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") on node \"crc\" " Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.756272 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-server-conf\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.756288 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.756302 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.756314 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bef00f62-0142-4a72-9849-12246c13d004-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.756325 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.756334 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bef00f62-0142-4a72-9849-12246c13d004-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.756345 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hhmz\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-kube-api-access-7hhmz\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.756356 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bef00f62-0142-4a72-9849-12246c13d004-pod-info\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.803734 4744 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.803907 4744 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-16439a32-bb72-4737-af92-d59e2f11ecb6" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6") on node "crc" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.808023 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "bef00f62-0142-4a72-9849-12246c13d004" (UID: "bef00f62-0142-4a72-9849-12246c13d004"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.859075 4744 reconciler_common.go:293] "Volume detached for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.859109 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bef00f62-0142-4a72-9849-12246c13d004-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:01 crc kubenswrapper[4744]: I0106 15:06:01.970753 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.436385 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.436381 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jndqv" event={"ID":"df07cf75-bdf5-43e2-8066-e1b4f2d7b583","Type":"ContainerStarted","Data":"0f07881fd6f2cf7fad444b13357e5f271c058d4fb714372c0b5007681855a1be"} Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.467449 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-jndqv" podStartSLOduration=15.421220772 podStartE2EDuration="45.467428409s" podCreationTimestamp="2026-01-06 15:05:17 +0000 UTC" firstStartedPulling="2026-01-06 15:05:31.919595919 +0000 UTC m=+1728.547062247" lastFinishedPulling="2026-01-06 15:06:01.965803556 +0000 UTC m=+1758.593269884" observedRunningTime="2026-01-06 15:06:02.456732726 +0000 UTC m=+1759.084199044" watchObservedRunningTime="2026-01-06 15:06:02.467428409 +0000 UTC m=+1759.094894727" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.497662 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.517788 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-1"] Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.533412 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Jan 06 15:06:02 crc kubenswrapper[4744]: E0106 15:06:02.534023 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7bf873c-7171-4ec5-a141-aeabfb5501df" containerName="heat-engine" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.534043 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7bf873c-7171-4ec5-a141-aeabfb5501df" containerName="heat-engine" Jan 06 15:06:02 crc kubenswrapper[4744]: E0106 15:06:02.534080 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="rabbitmq" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.534087 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="rabbitmq" Jan 06 15:06:02 crc kubenswrapper[4744]: E0106 15:06:02.534116 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="setup-container" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.534122 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="setup-container" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.534358 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef00f62-0142-4a72-9849-12246c13d004" containerName="rabbitmq" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.534381 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7bf873c-7171-4ec5-a141-aeabfb5501df" containerName="heat-engine" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.536252 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.546772 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.717647 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c38b64df-d42d-48ba-a5eb-12f0523a8555-config-data\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.717717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c38b64df-d42d-48ba-a5eb-12f0523a8555-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.717807 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdhfw\" (UniqueName: \"kubernetes.io/projected/c38b64df-d42d-48ba-a5eb-12f0523a8555-kube-api-access-pdhfw\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.717860 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c38b64df-d42d-48ba-a5eb-12f0523a8555-pod-info\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.717889 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.717933 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c38b64df-d42d-48ba-a5eb-12f0523a8555-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.717979 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.718003 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.718044 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.718066 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.718102 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c38b64df-d42d-48ba-a5eb-12f0523a8555-server-conf\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.820768 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdhfw\" (UniqueName: \"kubernetes.io/projected/c38b64df-d42d-48ba-a5eb-12f0523a8555-kube-api-access-pdhfw\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.820860 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c38b64df-d42d-48ba-a5eb-12f0523a8555-pod-info\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.820925 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.821007 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c38b64df-d42d-48ba-a5eb-12f0523a8555-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.821088 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.821108 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.821272 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.821336 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.821392 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c38b64df-d42d-48ba-a5eb-12f0523a8555-server-conf\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.821457 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c38b64df-d42d-48ba-a5eb-12f0523a8555-config-data\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.821506 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c38b64df-d42d-48ba-a5eb-12f0523a8555-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.822444 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c38b64df-d42d-48ba-a5eb-12f0523a8555-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.825312 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.828613 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c38b64df-d42d-48ba-a5eb-12f0523a8555-server-conf\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.829968 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c38b64df-d42d-48ba-a5eb-12f0523a8555-config-data\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.831825 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.831869 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d05509e9401e2174edbf89441304d15eb0beaf5fefd4c53cc6bfeba788bc2b73/globalmount\"" pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.833444 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.836307 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.837367 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c38b64df-d42d-48ba-a5eb-12f0523a8555-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.842377 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdhfw\" (UniqueName: \"kubernetes.io/projected/c38b64df-d42d-48ba-a5eb-12f0523a8555-kube-api-access-pdhfw\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.842622 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c38b64df-d42d-48ba-a5eb-12f0523a8555-pod-info\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.864852 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c38b64df-d42d-48ba-a5eb-12f0523a8555-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:02 crc kubenswrapper[4744]: I0106 15:06:02.947923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16439a32-bb72-4737-af92-d59e2f11ecb6\") pod \"rabbitmq-server-1\" (UID: \"c38b64df-d42d-48ba-a5eb-12f0523a8555\") " pod="openstack/rabbitmq-server-1" Jan 06 15:06:03 crc kubenswrapper[4744]: I0106 15:06:03.161416 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Jan 06 15:06:03 crc kubenswrapper[4744]: I0106 15:06:03.458537 4744 generic.go:334] "Generic (PLEG): container finished" podID="42eceb98-5fea-4ae7-827f-94b867781272" containerID="76d97059b01ffa66b999ea095cb76b4a2642e0a63f949bff76b983c119382f09" exitCode=0 Jan 06 15:06:03 crc kubenswrapper[4744]: I0106 15:06:03.458778 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" event={"ID":"42eceb98-5fea-4ae7-827f-94b867781272","Type":"ContainerDied","Data":"76d97059b01ffa66b999ea095cb76b4a2642e0a63f949bff76b983c119382f09"} Jan 06 15:06:03 crc kubenswrapper[4744]: I0106 15:06:03.731284 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bef00f62-0142-4a72-9849-12246c13d004" path="/var/lib/kubelet/pods/bef00f62-0142-4a72-9849-12246c13d004/volumes" Jan 06 15:06:03 crc kubenswrapper[4744]: I0106 15:06:03.892500 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Jan 06 15:06:04 crc kubenswrapper[4744]: I0106 15:06:04.475313 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"c38b64df-d42d-48ba-a5eb-12f0523a8555","Type":"ContainerStarted","Data":"b95869f145fd0c6366d2aa121704ca5f8fcc30065ce2ada7c3240bbb40142bb1"} Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.062811 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.223127 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-ssh-key-openstack-edpm-ipam\") pod \"42eceb98-5fea-4ae7-827f-94b867781272\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.223254 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-inventory\") pod \"42eceb98-5fea-4ae7-827f-94b867781272\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.223484 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6zzj\" (UniqueName: \"kubernetes.io/projected/42eceb98-5fea-4ae7-827f-94b867781272-kube-api-access-l6zzj\") pod \"42eceb98-5fea-4ae7-827f-94b867781272\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.223607 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-repo-setup-combined-ca-bundle\") pod \"42eceb98-5fea-4ae7-827f-94b867781272\" (UID: \"42eceb98-5fea-4ae7-827f-94b867781272\") " Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.229340 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "42eceb98-5fea-4ae7-827f-94b867781272" (UID: "42eceb98-5fea-4ae7-827f-94b867781272"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.229741 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42eceb98-5fea-4ae7-827f-94b867781272-kube-api-access-l6zzj" (OuterVolumeSpecName: "kube-api-access-l6zzj") pod "42eceb98-5fea-4ae7-827f-94b867781272" (UID: "42eceb98-5fea-4ae7-827f-94b867781272"). InnerVolumeSpecName "kube-api-access-l6zzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.261064 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "42eceb98-5fea-4ae7-827f-94b867781272" (UID: "42eceb98-5fea-4ae7-827f-94b867781272"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.274461 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-inventory" (OuterVolumeSpecName: "inventory") pod "42eceb98-5fea-4ae7-827f-94b867781272" (UID: "42eceb98-5fea-4ae7-827f-94b867781272"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.326056 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6zzj\" (UniqueName: \"kubernetes.io/projected/42eceb98-5fea-4ae7-827f-94b867781272-kube-api-access-l6zzj\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.326399 4744 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.326494 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.326585 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/42eceb98-5fea-4ae7-827f-94b867781272-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.499232 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" event={"ID":"42eceb98-5fea-4ae7-827f-94b867781272","Type":"ContainerDied","Data":"29cd3f48ede87559d88f380dafecb3fd6526d376aa736161ef95c53ddc31b2b4"} Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.499290 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29cd3f48ede87559d88f380dafecb3fd6526d376aa736161ef95c53ddc31b2b4" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.499288 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.619656 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w"] Jan 06 15:06:05 crc kubenswrapper[4744]: E0106 15:06:05.620460 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42eceb98-5fea-4ae7-827f-94b867781272" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.620493 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="42eceb98-5fea-4ae7-827f-94b867781272" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.620879 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="42eceb98-5fea-4ae7-827f-94b867781272" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.622105 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.625619 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.625982 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.626115 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.626212 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.644051 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w"] Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.760179 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65b52\" (UniqueName: \"kubernetes.io/projected/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-kube-api-access-65b52\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-x5b5w\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.760945 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-x5b5w\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.761102 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-x5b5w\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.865468 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-x5b5w\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.865555 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-x5b5w\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.865914 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65b52\" (UniqueName: \"kubernetes.io/projected/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-kube-api-access-65b52\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-x5b5w\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.873463 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-x5b5w\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.878068 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-x5b5w\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.890292 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65b52\" (UniqueName: \"kubernetes.io/projected/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-kube-api-access-65b52\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-x5b5w\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:05 crc kubenswrapper[4744]: I0106 15:06:05.958114 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:06 crc kubenswrapper[4744]: I0106 15:06:06.512375 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"c38b64df-d42d-48ba-a5eb-12f0523a8555","Type":"ContainerStarted","Data":"771a6d362099a5d73a7f5b99bf51c54a26ae8fa8fdf102c1e8b70441a183cabd"} Jan 06 15:06:06 crc kubenswrapper[4744]: I0106 15:06:06.516835 4744 generic.go:334] "Generic (PLEG): container finished" podID="df07cf75-bdf5-43e2-8066-e1b4f2d7b583" containerID="0f07881fd6f2cf7fad444b13357e5f271c058d4fb714372c0b5007681855a1be" exitCode=0 Jan 06 15:06:06 crc kubenswrapper[4744]: I0106 15:06:06.516875 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jndqv" event={"ID":"df07cf75-bdf5-43e2-8066-e1b4f2d7b583","Type":"ContainerDied","Data":"0f07881fd6f2cf7fad444b13357e5f271c058d4fb714372c0b5007681855a1be"} Jan 06 15:06:06 crc kubenswrapper[4744]: I0106 15:06:06.645331 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w"] Jan 06 15:06:07 crc kubenswrapper[4744]: I0106 15:06:07.527740 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" event={"ID":"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76","Type":"ContainerStarted","Data":"8fef5c8e2b626e7bf72af5140f55ab7fbc0e026cb0600dd400d52c63423e5e4e"} Jan 06 15:06:07 crc kubenswrapper[4744]: I0106 15:06:07.899023 4744 scope.go:117] "RemoveContainer" containerID="738a503b4e10e99e3abe518b1919d9ce5bac03ecfcbea228a9694a955f0cf80d" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.002565 4744 scope.go:117] "RemoveContainer" containerID="98ff97c26acef954ef996e4aacd4e7dec375b11833016c6e200cd86dd7f206d4" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.051703 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jndqv" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.135129 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-combined-ca-bundle\") pod \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.135259 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-config-data\") pod \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.135369 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fctmc\" (UniqueName: \"kubernetes.io/projected/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-kube-api-access-fctmc\") pod \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.135531 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-scripts\") pod \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\" (UID: \"df07cf75-bdf5-43e2-8066-e1b4f2d7b583\") " Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.141278 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-kube-api-access-fctmc" (OuterVolumeSpecName: "kube-api-access-fctmc") pod "df07cf75-bdf5-43e2-8066-e1b4f2d7b583" (UID: "df07cf75-bdf5-43e2-8066-e1b4f2d7b583"). InnerVolumeSpecName "kube-api-access-fctmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.143753 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-scripts" (OuterVolumeSpecName: "scripts") pod "df07cf75-bdf5-43e2-8066-e1b4f2d7b583" (UID: "df07cf75-bdf5-43e2-8066-e1b4f2d7b583"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.171547 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df07cf75-bdf5-43e2-8066-e1b4f2d7b583" (UID: "df07cf75-bdf5-43e2-8066-e1b4f2d7b583"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.176726 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-config-data" (OuterVolumeSpecName: "config-data") pod "df07cf75-bdf5-43e2-8066-e1b4f2d7b583" (UID: "df07cf75-bdf5-43e2-8066-e1b4f2d7b583"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.239299 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.239344 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.239359 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.239373 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fctmc\" (UniqueName: \"kubernetes.io/projected/df07cf75-bdf5-43e2-8066-e1b4f2d7b583-kube-api-access-fctmc\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.539456 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-jndqv" event={"ID":"df07cf75-bdf5-43e2-8066-e1b4f2d7b583","Type":"ContainerDied","Data":"3d9cf0650e7d0fe5576b66380eca3260d86fd574a84ee2941b4ffb3c55920887"} Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.540690 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d9cf0650e7d0fe5576b66380eca3260d86fd574a84ee2941b4ffb3c55920887" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.539481 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-jndqv" Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.541089 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" event={"ID":"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76","Type":"ContainerStarted","Data":"82b24539b891efba0eb8f002cbca09739436f481f19fbbe103ad0466352ffd6c"} Jan 06 15:06:08 crc kubenswrapper[4744]: I0106 15:06:08.575390 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" podStartSLOduration=3.062331128 podStartE2EDuration="3.575361744s" podCreationTimestamp="2026-01-06 15:06:05 +0000 UTC" firstStartedPulling="2026-01-06 15:06:06.654580127 +0000 UTC m=+1763.282046445" lastFinishedPulling="2026-01-06 15:06:07.167610743 +0000 UTC m=+1763.795077061" observedRunningTime="2026-01-06 15:06:08.561295242 +0000 UTC m=+1765.188761590" watchObservedRunningTime="2026-01-06 15:06:08.575361744 +0000 UTC m=+1765.202828082" Jan 06 15:06:09 crc kubenswrapper[4744]: I0106 15:06:09.712725 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:06:09 crc kubenswrapper[4744]: E0106 15:06:09.713283 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:06:11 crc kubenswrapper[4744]: I0106 15:06:11.576320 4744 generic.go:334] "Generic (PLEG): container finished" podID="8664eee2-4c8b-4e3c-9bdd-26658aa1ab76" containerID="82b24539b891efba0eb8f002cbca09739436f481f19fbbe103ad0466352ffd6c" exitCode=0 Jan 06 15:06:11 crc kubenswrapper[4744]: I0106 15:06:11.576416 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" event={"ID":"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76","Type":"ContainerDied","Data":"82b24539b891efba0eb8f002cbca09739436f481f19fbbe103ad0466352ffd6c"} Jan 06 15:06:12 crc kubenswrapper[4744]: I0106 15:06:12.483447 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Jan 06 15:06:12 crc kubenswrapper[4744]: I0106 15:06:12.484073 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-api" containerID="cri-o://f6abcbfa1d6d72ebb5e38535ffbbe20cedbd190b678474b2ddc6155c7750908d" gracePeriod=30 Jan 06 15:06:12 crc kubenswrapper[4744]: I0106 15:06:12.484297 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-evaluator" containerID="cri-o://f4e4ab41eea7a7ea2b747213b9ef661ad9bb1c9706f988b8b509313c22fe3621" gracePeriod=30 Jan 06 15:06:12 crc kubenswrapper[4744]: I0106 15:06:12.484297 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-notifier" containerID="cri-o://223c1c237ab1a8c64b3f384246a510c319b92d8cf96efdcaa0c4ca3a382cccd3" gracePeriod=30 Jan 06 15:06:12 crc kubenswrapper[4744]: I0106 15:06:12.484354 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-listener" containerID="cri-o://f6ef551a84bf8d151ee9b9d821dc1ba8fa2613e0e6488a9483a42deaedfa40f5" gracePeriod=30 Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.192993 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.287840 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65b52\" (UniqueName: \"kubernetes.io/projected/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-kube-api-access-65b52\") pod \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.288074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-ssh-key-openstack-edpm-ipam\") pod \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.288099 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-inventory\") pod \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\" (UID: \"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76\") " Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.298537 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-kube-api-access-65b52" (OuterVolumeSpecName: "kube-api-access-65b52") pod "8664eee2-4c8b-4e3c-9bdd-26658aa1ab76" (UID: "8664eee2-4c8b-4e3c-9bdd-26658aa1ab76"). InnerVolumeSpecName "kube-api-access-65b52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.324260 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8664eee2-4c8b-4e3c-9bdd-26658aa1ab76" (UID: "8664eee2-4c8b-4e3c-9bdd-26658aa1ab76"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.350080 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-inventory" (OuterVolumeSpecName: "inventory") pod "8664eee2-4c8b-4e3c-9bdd-26658aa1ab76" (UID: "8664eee2-4c8b-4e3c-9bdd-26658aa1ab76"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.390377 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65b52\" (UniqueName: \"kubernetes.io/projected/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-kube-api-access-65b52\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.390413 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.390425 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8664eee2-4c8b-4e3c-9bdd-26658aa1ab76-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:13 crc kubenswrapper[4744]: E0106 15:06:13.503072 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85034f0d_f575_44da_9f68_9d5c2c2e6f74.slice/crio-conmon-f4e4ab41eea7a7ea2b747213b9ef661ad9bb1c9706f988b8b509313c22fe3621.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.602875 4744 generic.go:334] "Generic (PLEG): container finished" podID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerID="f4e4ab41eea7a7ea2b747213b9ef661ad9bb1c9706f988b8b509313c22fe3621" exitCode=0 Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.602916 4744 generic.go:334] "Generic (PLEG): container finished" podID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerID="f6abcbfa1d6d72ebb5e38535ffbbe20cedbd190b678474b2ddc6155c7750908d" exitCode=0 Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.602914 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerDied","Data":"f4e4ab41eea7a7ea2b747213b9ef661ad9bb1c9706f988b8b509313c22fe3621"} Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.602974 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerDied","Data":"f6abcbfa1d6d72ebb5e38535ffbbe20cedbd190b678474b2ddc6155c7750908d"} Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.605135 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" event={"ID":"8664eee2-4c8b-4e3c-9bdd-26658aa1ab76","Type":"ContainerDied","Data":"8fef5c8e2b626e7bf72af5140f55ab7fbc0e026cb0600dd400d52c63423e5e4e"} Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.605195 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fef5c8e2b626e7bf72af5140f55ab7fbc0e026cb0600dd400d52c63423e5e4e" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.605261 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-x5b5w" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.694799 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz"] Jan 06 15:06:13 crc kubenswrapper[4744]: E0106 15:06:13.695388 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8664eee2-4c8b-4e3c-9bdd-26658aa1ab76" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.695408 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8664eee2-4c8b-4e3c-9bdd-26658aa1ab76" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 06 15:06:13 crc kubenswrapper[4744]: E0106 15:06:13.695460 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df07cf75-bdf5-43e2-8066-e1b4f2d7b583" containerName="aodh-db-sync" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.695467 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="df07cf75-bdf5-43e2-8066-e1b4f2d7b583" containerName="aodh-db-sync" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.695693 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="df07cf75-bdf5-43e2-8066-e1b4f2d7b583" containerName="aodh-db-sync" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.695720 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8664eee2-4c8b-4e3c-9bdd-26658aa1ab76" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.696575 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.698745 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.699406 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.699975 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.700289 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.745302 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz"] Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.800083 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.800598 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.801384 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zngx9\" (UniqueName: \"kubernetes.io/projected/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-kube-api-access-zngx9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.801555 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.916889 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.917080 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.917224 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zngx9\" (UniqueName: \"kubernetes.io/projected/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-kube-api-access-zngx9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.917340 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.922106 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.922215 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.925863 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:13 crc kubenswrapper[4744]: I0106 15:06:13.941809 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zngx9\" (UniqueName: \"kubernetes.io/projected/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-kube-api-access-zngx9\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:14 crc kubenswrapper[4744]: I0106 15:06:14.020711 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:06:15 crc kubenswrapper[4744]: I0106 15:06:15.577037 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz"] Jan 06 15:06:15 crc kubenswrapper[4744]: I0106 15:06:15.917868 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" event={"ID":"eac03a71-e859-4391-b58e-f4dc4fdfa0ce","Type":"ContainerStarted","Data":"b5496902710052f46c0aa76fd721cdd07542a479ee5e3e740144b650244b0a6f"} Jan 06 15:06:15 crc kubenswrapper[4744]: I0106 15:06:15.920850 4744 generic.go:334] "Generic (PLEG): container finished" podID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerID="f6ef551a84bf8d151ee9b9d821dc1ba8fa2613e0e6488a9483a42deaedfa40f5" exitCode=0 Jan 06 15:06:15 crc kubenswrapper[4744]: I0106 15:06:15.920878 4744 generic.go:334] "Generic (PLEG): container finished" podID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerID="223c1c237ab1a8c64b3f384246a510c319b92d8cf96efdcaa0c4ca3a382cccd3" exitCode=0 Jan 06 15:06:15 crc kubenswrapper[4744]: I0106 15:06:15.920892 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerDied","Data":"f6ef551a84bf8d151ee9b9d821dc1ba8fa2613e0e6488a9483a42deaedfa40f5"} Jan 06 15:06:15 crc kubenswrapper[4744]: I0106 15:06:15.920910 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerDied","Data":"223c1c237ab1a8c64b3f384246a510c319b92d8cf96efdcaa0c4ca3a382cccd3"} Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.167978 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.315459 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-combined-ca-bundle\") pod \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.315804 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-public-tls-certs\") pod \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.316010 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-config-data\") pod \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.316285 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-scripts\") pod \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.316512 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-internal-tls-certs\") pod \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.316893 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwphb\" (UniqueName: \"kubernetes.io/projected/85034f0d-f575-44da-9f68-9d5c2c2e6f74-kube-api-access-kwphb\") pod \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\" (UID: \"85034f0d-f575-44da-9f68-9d5c2c2e6f74\") " Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.322576 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-scripts" (OuterVolumeSpecName: "scripts") pod "85034f0d-f575-44da-9f68-9d5c2c2e6f74" (UID: "85034f0d-f575-44da-9f68-9d5c2c2e6f74"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.334405 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85034f0d-f575-44da-9f68-9d5c2c2e6f74-kube-api-access-kwphb" (OuterVolumeSpecName: "kube-api-access-kwphb") pod "85034f0d-f575-44da-9f68-9d5c2c2e6f74" (UID: "85034f0d-f575-44da-9f68-9d5c2c2e6f74"). InnerVolumeSpecName "kube-api-access-kwphb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.399150 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "85034f0d-f575-44da-9f68-9d5c2c2e6f74" (UID: "85034f0d-f575-44da-9f68-9d5c2c2e6f74"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.416878 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "85034f0d-f575-44da-9f68-9d5c2c2e6f74" (UID: "85034f0d-f575-44da-9f68-9d5c2c2e6f74"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.422470 4744 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-scripts\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.422527 4744 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.422542 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwphb\" (UniqueName: \"kubernetes.io/projected/85034f0d-f575-44da-9f68-9d5c2c2e6f74-kube-api-access-kwphb\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.422551 4744 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.464854 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-config-data" (OuterVolumeSpecName: "config-data") pod "85034f0d-f575-44da-9f68-9d5c2c2e6f74" (UID: "85034f0d-f575-44da-9f68-9d5c2c2e6f74"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.479143 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85034f0d-f575-44da-9f68-9d5c2c2e6f74" (UID: "85034f0d-f575-44da-9f68-9d5c2c2e6f74"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.524986 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.525041 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85034f0d-f575-44da-9f68-9d5c2c2e6f74-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.934967 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" event={"ID":"eac03a71-e859-4391-b58e-f4dc4fdfa0ce","Type":"ContainerStarted","Data":"d932e43f1febe26a75c11afe102a8d6bd107eca7366410e9cecb37db8f233871"} Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.938119 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"85034f0d-f575-44da-9f68-9d5c2c2e6f74","Type":"ContainerDied","Data":"54bc70450c031312f7989a99a4633cfa42fe6e015473731ac24ef8655ea017fb"} Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.938216 4744 scope.go:117] "RemoveContainer" containerID="f6ef551a84bf8d151ee9b9d821dc1ba8fa2613e0e6488a9483a42deaedfa40f5" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.938217 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.968433 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" podStartSLOduration=3.515263065 podStartE2EDuration="3.968404546s" podCreationTimestamp="2026-01-06 15:06:13 +0000 UTC" firstStartedPulling="2026-01-06 15:06:15.569294914 +0000 UTC m=+1772.196761232" lastFinishedPulling="2026-01-06 15:06:16.022436395 +0000 UTC m=+1772.649902713" observedRunningTime="2026-01-06 15:06:16.952148455 +0000 UTC m=+1773.579614773" watchObservedRunningTime="2026-01-06 15:06:16.968404546 +0000 UTC m=+1773.595870864" Jan 06 15:06:16 crc kubenswrapper[4744]: I0106 15:06:16.984750 4744 scope.go:117] "RemoveContainer" containerID="223c1c237ab1a8c64b3f384246a510c319b92d8cf96efdcaa0c4ca3a382cccd3" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.003031 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.018559 4744 scope.go:117] "RemoveContainer" containerID="f4e4ab41eea7a7ea2b747213b9ef661ad9bb1c9706f988b8b509313c22fe3621" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.019322 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.037593 4744 scope.go:117] "RemoveContainer" containerID="f6abcbfa1d6d72ebb5e38535ffbbe20cedbd190b678474b2ddc6155c7750908d" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.039430 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Jan 06 15:06:17 crc kubenswrapper[4744]: E0106 15:06:17.040090 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-listener" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.040121 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-listener" Jan 06 15:06:17 crc kubenswrapper[4744]: E0106 15:06:17.040140 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-evaluator" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.040149 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-evaluator" Jan 06 15:06:17 crc kubenswrapper[4744]: E0106 15:06:17.040183 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-notifier" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.040193 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-notifier" Jan 06 15:06:17 crc kubenswrapper[4744]: E0106 15:06:17.040233 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-api" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.040241 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-api" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.040568 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-api" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.040595 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-notifier" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.040624 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-evaluator" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.040645 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" containerName="aodh-listener" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.047863 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.049787 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.051686 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.051900 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.052218 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-tjbk8" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.052265 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.053372 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.151040 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d699\" (UniqueName: \"kubernetes.io/projected/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-kube-api-access-6d699\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.151208 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-internal-tls-certs\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.151257 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.151312 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-scripts\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.151360 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-public-tls-certs\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.151397 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-config-data\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.252230 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.252301 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-scripts\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.252341 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-public-tls-certs\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.252367 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-config-data\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.252446 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d699\" (UniqueName: \"kubernetes.io/projected/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-kube-api-access-6d699\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.252516 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-internal-tls-certs\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.257048 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-public-tls-certs\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.257586 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-config-data\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.258125 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-scripts\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.258991 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.269809 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-internal-tls-certs\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.270985 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d699\" (UniqueName: \"kubernetes.io/projected/d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f-kube-api-access-6d699\") pod \"aodh-0\" (UID: \"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f\") " pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.376685 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.730174 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85034f0d-f575-44da-9f68-9d5c2c2e6f74" path="/var/lib/kubelet/pods/85034f0d-f575-44da-9f68-9d5c2c2e6f74/volumes" Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.885847 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Jan 06 15:06:17 crc kubenswrapper[4744]: W0106 15:06:17.901954 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd77ef1d6_bd13_43f6_ac08_3da1f1a20d4f.slice/crio-fa255ac8ee481402cab94bc64814675fbb2c5f99af3fa444930502a35101af27 WatchSource:0}: Error finding container fa255ac8ee481402cab94bc64814675fbb2c5f99af3fa444930502a35101af27: Status 404 returned error can't find the container with id fa255ac8ee481402cab94bc64814675fbb2c5f99af3fa444930502a35101af27 Jan 06 15:06:17 crc kubenswrapper[4744]: I0106 15:06:17.952715 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f","Type":"ContainerStarted","Data":"fa255ac8ee481402cab94bc64814675fbb2c5f99af3fa444930502a35101af27"} Jan 06 15:06:18 crc kubenswrapper[4744]: I0106 15:06:18.973071 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f","Type":"ContainerStarted","Data":"311e0e3ea186a15ac58b084a24b1c8041dc4e105eab931a52a323b7faa2090bb"} Jan 06 15:06:19 crc kubenswrapper[4744]: I0106 15:06:19.987107 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f","Type":"ContainerStarted","Data":"4cd28b9775069064c584aa60d3af5841d1c2037332fa94bc6db37f656274f497"} Jan 06 15:06:22 crc kubenswrapper[4744]: I0106 15:06:22.023393 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f","Type":"ContainerStarted","Data":"4317d8942517768154005df2deeb96fe1c9860b852dce92b3008c6b8206ed9f2"} Jan 06 15:06:23 crc kubenswrapper[4744]: I0106 15:06:23.039285 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f","Type":"ContainerStarted","Data":"3264edbf7399e18e010fd7cc486bc21a0dda2e0e2a77284ff9db82f91eecdc67"} Jan 06 15:06:23 crc kubenswrapper[4744]: I0106 15:06:23.069391 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.968553004 podStartE2EDuration="7.069365877s" podCreationTimestamp="2026-01-06 15:06:16 +0000 UTC" firstStartedPulling="2026-01-06 15:06:17.906170271 +0000 UTC m=+1774.533636589" lastFinishedPulling="2026-01-06 15:06:22.006983144 +0000 UTC m=+1778.634449462" observedRunningTime="2026-01-06 15:06:23.065953706 +0000 UTC m=+1779.693420044" watchObservedRunningTime="2026-01-06 15:06:23.069365877 +0000 UTC m=+1779.696832195" Jan 06 15:06:24 crc kubenswrapper[4744]: I0106 15:06:24.712375 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:06:24 crc kubenswrapper[4744]: E0106 15:06:24.714124 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:06:39 crc kubenswrapper[4744]: I0106 15:06:39.711817 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:06:39 crc kubenswrapper[4744]: E0106 15:06:39.712953 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:06:39 crc kubenswrapper[4744]: I0106 15:06:39.979006 4744 patch_prober.go:28] interesting pod/monitoring-plugin-8468c6fcd4-8kpwb container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.83:9443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 15:06:39 crc kubenswrapper[4744]: I0106 15:06:39.979092 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-8468c6fcd4-8kpwb" podUID="9149e4db-4b7e-4247-ad40-5b2b9738aa38" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.83:9443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 06 15:06:44 crc kubenswrapper[4744]: I0106 15:06:44.336067 4744 generic.go:334] "Generic (PLEG): container finished" podID="c38b64df-d42d-48ba-a5eb-12f0523a8555" containerID="771a6d362099a5d73a7f5b99bf51c54a26ae8fa8fdf102c1e8b70441a183cabd" exitCode=0 Jan 06 15:06:44 crc kubenswrapper[4744]: I0106 15:06:44.336226 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"c38b64df-d42d-48ba-a5eb-12f0523a8555","Type":"ContainerDied","Data":"771a6d362099a5d73a7f5b99bf51c54a26ae8fa8fdf102c1e8b70441a183cabd"} Jan 06 15:06:45 crc kubenswrapper[4744]: I0106 15:06:45.351070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"c38b64df-d42d-48ba-a5eb-12f0523a8555","Type":"ContainerStarted","Data":"c5b708098042d59ca0dce26d39498a0396997492f667b45ceb3a06700d98a67d"} Jan 06 15:06:45 crc kubenswrapper[4744]: I0106 15:06:45.352096 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Jan 06 15:06:45 crc kubenswrapper[4744]: I0106 15:06:45.389895 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=43.389874709 podStartE2EDuration="43.389874709s" podCreationTimestamp="2026-01-06 15:06:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:06:45.374498582 +0000 UTC m=+1802.001964910" watchObservedRunningTime="2026-01-06 15:06:45.389874709 +0000 UTC m=+1802.017341027" Jan 06 15:06:54 crc kubenswrapper[4744]: I0106 15:06:54.712035 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:06:55 crc kubenswrapper[4744]: I0106 15:06:55.474735 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"8d70e5126160c45c7349d68db9fd09dd06c15ea7a0105b790d710f8a5423e632"} Jan 06 15:07:03 crc kubenswrapper[4744]: I0106 15:07:03.163331 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Jan 06 15:07:03 crc kubenswrapper[4744]: I0106 15:07:03.230412 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 06 15:07:07 crc kubenswrapper[4744]: I0106 15:07:07.381303 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="rabbitmq" containerID="cri-o://c6f4211a18370302a96e5b9028a1ccfd0b2c832ac6b5ed9dbcf6df60c491d43f" gracePeriod=604796 Jan 06 15:07:08 crc kubenswrapper[4744]: I0106 15:07:08.220046 4744 scope.go:117] "RemoveContainer" containerID="2ab1b4a19c367c9ffceebc495e6f631a1fd7c02935d8396230ea56f4d9c7162f" Jan 06 15:07:08 crc kubenswrapper[4744]: I0106 15:07:08.262609 4744 scope.go:117] "RemoveContainer" containerID="d365e3d5aee297c5b1288d55446a35840c4d302794f56ad081a607472297daf8" Jan 06 15:07:08 crc kubenswrapper[4744]: I0106 15:07:08.308053 4744 scope.go:117] "RemoveContainer" containerID="845cc2c773efac6f0e9deb2039d627164a3490782ca8e57484b4b28d025efc01" Jan 06 15:07:08 crc kubenswrapper[4744]: I0106 15:07:08.372823 4744 scope.go:117] "RemoveContainer" containerID="f99ed76641c6b5de4c93710c95b2260d76b028d6b87f58fff59251657ee737fb" Jan 06 15:07:08 crc kubenswrapper[4744]: I0106 15:07:08.394818 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.128:5671: connect: connection refused" Jan 06 15:07:08 crc kubenswrapper[4744]: I0106 15:07:08.429216 4744 scope.go:117] "RemoveContainer" containerID="3a30acde6ec885c6d3d45d8d54558bc2d1b46b6c50bfdfb264142add4bc8064a" Jan 06 15:07:12 crc kubenswrapper[4744]: I0106 15:07:12.079036 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-2rcd6"] Jan 06 15:07:12 crc kubenswrapper[4744]: I0106 15:07:12.099009 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-2rcd6"] Jan 06 15:07:13 crc kubenswrapper[4744]: I0106 15:07:13.033078 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-bv5rj"] Jan 06 15:07:13 crc kubenswrapper[4744]: I0106 15:07:13.048101 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-bv5rj"] Jan 06 15:07:13 crc kubenswrapper[4744]: I0106 15:07:13.735621 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf899aa6-5d99-482f-b493-56d7de87e44c" path="/var/lib/kubelet/pods/bf899aa6-5d99-482f-b493-56d7de87e44c/volumes" Jan 06 15:07:13 crc kubenswrapper[4744]: I0106 15:07:13.807629 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f410252c-e5a3-4d42-ba06-febebef5a30a" path="/var/lib/kubelet/pods/f410252c-e5a3-4d42-ba06-febebef5a30a/volumes" Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.057042 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-6c8a-account-create-update-7b4pf"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.075101 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-6c8a-account-create-update-7b4pf"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.091626 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8d3a-account-create-update-fhjdm"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.101319 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-thfs7"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.112268 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8d3a-account-create-update-fhjdm"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.124532 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-79d0-account-create-update-w29fr"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.138454 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f0f3-account-create-update-h8jq2"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.151821 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-79d0-account-create-update-w29fr"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.169142 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-thfs7"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.179773 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-f0f3-account-create-update-h8jq2"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.190112 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-x5r7t"] Jan 06 15:07:14 crc kubenswrapper[4744]: I0106 15:07:14.199886 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-x5r7t"] Jan 06 15:07:15 crc kubenswrapper[4744]: I0106 15:07:15.730142 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="218cb457-4148-4612-bbf6-f8240c136d3b" path="/var/lib/kubelet/pods/218cb457-4148-4612-bbf6-f8240c136d3b/volumes" Jan 06 15:07:15 crc kubenswrapper[4744]: I0106 15:07:15.733472 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4136bc47-2e87-4af7-acf9-23a36c37269d" path="/var/lib/kubelet/pods/4136bc47-2e87-4af7-acf9-23a36c37269d/volumes" Jan 06 15:07:15 crc kubenswrapper[4744]: I0106 15:07:15.734323 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c10fd027-3c56-4a1d-a16f-de6b641a82f4" path="/var/lib/kubelet/pods/c10fd027-3c56-4a1d-a16f-de6b641a82f4/volumes" Jan 06 15:07:15 crc kubenswrapper[4744]: I0106 15:07:15.735350 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d755b346-e89a-4ae8-a162-4fc6f8c28797" path="/var/lib/kubelet/pods/d755b346-e89a-4ae8-a162-4fc6f8c28797/volumes" Jan 06 15:07:15 crc kubenswrapper[4744]: I0106 15:07:15.737987 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e8cff6-96c6-4fee-99ca-1f77403080b4" path="/var/lib/kubelet/pods/d7e8cff6-96c6-4fee-99ca-1f77403080b4/volumes" Jan 06 15:07:15 crc kubenswrapper[4744]: I0106 15:07:15.739557 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de589768-1fff-4511-a85e-cbb4e8cb30ed" path="/var/lib/kubelet/pods/de589768-1fff-4511-a85e-cbb4e8cb30ed/volumes" Jan 06 15:07:15 crc kubenswrapper[4744]: I0106 15:07:15.813481 4744 generic.go:334] "Generic (PLEG): container finished" podID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerID="c6f4211a18370302a96e5b9028a1ccfd0b2c832ac6b5ed9dbcf6df60c491d43f" exitCode=0 Jan 06 15:07:15 crc kubenswrapper[4744]: I0106 15:07:15.813892 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff","Type":"ContainerDied","Data":"c6f4211a18370302a96e5b9028a1ccfd0b2c832ac6b5ed9dbcf6df60c491d43f"} Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.099235 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.169571 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-erlang-cookie\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.169683 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-erlang-cookie-secret\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.169716 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-plugins\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.169752 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-server-conf\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.169777 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-pod-info\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.169867 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-plugins-conf\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.171467 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.171567 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-confd\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.171623 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4jdt\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-kube-api-access-f4jdt\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.171865 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-tls\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.171918 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-config-data\") pod \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\" (UID: \"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff\") " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.176584 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.179510 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.187979 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-kube-api-access-f4jdt" (OuterVolumeSpecName: "kube-api-access-f4jdt") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "kube-api-access-f4jdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.190341 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-pod-info" (OuterVolumeSpecName: "pod-info") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.202068 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.202534 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.209374 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.228534 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127" (OuterVolumeSpecName: "persistence") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "pvc-ed0a5b62-b676-45f3-a80e-b9de14126127". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.238739 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-config-data" (OuterVolumeSpecName: "config-data") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.275404 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.275439 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.275449 4744 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.275458 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.275466 4744 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-pod-info\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.275477 4744 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.275507 4744 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") on node \"crc\" " Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.275517 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4jdt\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-kube-api-access-f4jdt\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.275527 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.297755 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-server-conf" (OuterVolumeSpecName: "server-conf") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.311129 4744 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.311284 4744 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ed0a5b62-b676-45f3-a80e-b9de14126127" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127") on node "crc" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.355323 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" (UID: "f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.377841 4744 reconciler_common.go:293] "Volume detached for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.377875 4744 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.377889 4744 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff-server-conf\") on node \"crc\" DevicePath \"\"" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.826892 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff","Type":"ContainerDied","Data":"43318afe9772285af4911e794bb68f4d82640a7d7352ed2508bcc0eafb7237a1"} Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.827272 4744 scope.go:117] "RemoveContainer" containerID="c6f4211a18370302a96e5b9028a1ccfd0b2c832ac6b5ed9dbcf6df60c491d43f" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.827290 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.858796 4744 scope.go:117] "RemoveContainer" containerID="b568e5b0f4450110f1563bd5f836d3c37170139a7c49b77607d38119a536ea25" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.880274 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.899958 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.928272 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 06 15:07:16 crc kubenswrapper[4744]: E0106 15:07:16.928713 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="setup-container" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.928731 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="setup-container" Jan 06 15:07:16 crc kubenswrapper[4744]: E0106 15:07:16.928751 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="rabbitmq" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.928757 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="rabbitmq" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.928989 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" containerName="rabbitmq" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.931897 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 06 15:07:16 crc kubenswrapper[4744]: I0106 15:07:16.954141 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099126 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099343 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4c6a122-8746-4f2b-a7f9-e3df62969170-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099391 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099419 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099454 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099493 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4c6a122-8746-4f2b-a7f9-e3df62969170-config-data\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099528 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4c6a122-8746-4f2b-a7f9-e3df62969170-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099559 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4c6a122-8746-4f2b-a7f9-e3df62969170-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099667 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099704 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4c6a122-8746-4f2b-a7f9-e3df62969170-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.099731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9bnh\" (UniqueName: \"kubernetes.io/projected/b4c6a122-8746-4f2b-a7f9-e3df62969170-kube-api-access-m9bnh\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.202513 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.202574 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.202613 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.202656 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4c6a122-8746-4f2b-a7f9-e3df62969170-config-data\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.202691 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4c6a122-8746-4f2b-a7f9-e3df62969170-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.202719 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4c6a122-8746-4f2b-a7f9-e3df62969170-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.202804 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.202837 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4c6a122-8746-4f2b-a7f9-e3df62969170-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.202864 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9bnh\" (UniqueName: \"kubernetes.io/projected/b4c6a122-8746-4f2b-a7f9-e3df62969170-kube-api-access-m9bnh\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.203041 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.203086 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4c6a122-8746-4f2b-a7f9-e3df62969170-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.204252 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4c6a122-8746-4f2b-a7f9-e3df62969170-config-data\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.204642 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.204821 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.205016 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b4c6a122-8746-4f2b-a7f9-e3df62969170-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.205267 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b4c6a122-8746-4f2b-a7f9-e3df62969170-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.207382 4744 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.207551 4744 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4922a1e397587981e3404408ee9f34eec949eb29c3877b1b5adb0ae562e7c69f/globalmount\"" pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.210014 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b4c6a122-8746-4f2b-a7f9-e3df62969170-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.210035 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b4c6a122-8746-4f2b-a7f9-e3df62969170-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.210075 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.210333 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b4c6a122-8746-4f2b-a7f9-e3df62969170-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.225723 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9bnh\" (UniqueName: \"kubernetes.io/projected/b4c6a122-8746-4f2b-a7f9-e3df62969170-kube-api-access-m9bnh\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.284543 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ed0a5b62-b676-45f3-a80e-b9de14126127\") pod \"rabbitmq-server-0\" (UID: \"b4c6a122-8746-4f2b-a7f9-e3df62969170\") " pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.588095 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 06 15:07:17 crc kubenswrapper[4744]: I0106 15:07:17.754680 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff" path="/var/lib/kubelet/pods/f2ae7ac8-c73e-4cf2-aa51-3c6f7c47e1ff/volumes" Jan 06 15:07:18 crc kubenswrapper[4744]: I0106 15:07:18.142065 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 06 15:07:18 crc kubenswrapper[4744]: I0106 15:07:18.876990 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b4c6a122-8746-4f2b-a7f9-e3df62969170","Type":"ContainerStarted","Data":"4b24185c7155826c9ebdd21e064de019920734e65b5e97f15e32fa21e27c0c90"} Jan 06 15:07:21 crc kubenswrapper[4744]: I0106 15:07:21.919548 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b4c6a122-8746-4f2b-a7f9-e3df62969170","Type":"ContainerStarted","Data":"aa505f7cef8367ccdaffa0f7a55a505803b66e35822b89aa6f786905ee0543b2"} Jan 06 15:07:23 crc kubenswrapper[4744]: I0106 15:07:23.035143 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf"] Jan 06 15:07:23 crc kubenswrapper[4744]: I0106 15:07:23.052019 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-sfjpf"] Jan 06 15:07:23 crc kubenswrapper[4744]: I0106 15:07:23.728999 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d656c5-0291-4edd-b1cd-304b1bdce02b" path="/var/lib/kubelet/pods/67d656c5-0291-4edd-b1cd-304b1bdce02b/volumes" Jan 06 15:07:26 crc kubenswrapper[4744]: I0106 15:07:26.045535 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-08dd-account-create-update-hph9l"] Jan 06 15:07:26 crc kubenswrapper[4744]: I0106 15:07:26.057488 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-8lng4"] Jan 06 15:07:26 crc kubenswrapper[4744]: I0106 15:07:26.068243 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-08dd-account-create-update-hph9l"] Jan 06 15:07:26 crc kubenswrapper[4744]: I0106 15:07:26.082068 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-8lng4"] Jan 06 15:07:27 crc kubenswrapper[4744]: I0106 15:07:27.734411 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5d7e690-64ff-4302-ab9d-57685f2abc88" path="/var/lib/kubelet/pods/b5d7e690-64ff-4302-ab9d-57685f2abc88/volumes" Jan 06 15:07:27 crc kubenswrapper[4744]: I0106 15:07:27.735780 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f59d0597-674a-4db8-b70c-1307186cf198" path="/var/lib/kubelet/pods/f59d0597-674a-4db8-b70c-1307186cf198/volumes" Jan 06 15:07:29 crc kubenswrapper[4744]: I0106 15:07:29.166506 4744 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-m2bjv container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 15:07:29 crc kubenswrapper[4744]: I0106 15:07:29.166934 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m2bjv" podUID="b3bf455e-520f-439a-a558-5f6b1feb046b" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:07:31 crc kubenswrapper[4744]: I0106 15:07:31.385844 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" podUID="2bf34d55-4b6d-4716-aca0-026cdd02aabe" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 15:07:31 crc kubenswrapper[4744]: I0106 15:07:31.418500 4744 trace.go:236] Trace[294893396]: "Calculate volume metrics of prometheus-metric-storage-db for pod openstack/prometheus-metric-storage-0" (06-Jan-2026 15:07:28.521) (total time: 2896ms): Jan 06 15:07:31 crc kubenswrapper[4744]: Trace[294893396]: [2.896492279s] [2.896492279s] END Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.065670 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-79c64"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.076601 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f033-account-create-update-w89dh"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.089980 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-f5a5-account-create-update-jkh64"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.099707 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-6252-account-create-update-cz7bz"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.108377 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-gwmvk"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.117324 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-67ffc"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.125695 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-s8j7d"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.141734 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6b41-account-create-update-9qd64"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.155362 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-f5a5-account-create-update-jkh64"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.168969 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-79c64"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.178746 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f033-account-create-update-w89dh"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.187873 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6b41-account-create-update-9qd64"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.197363 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-s8j7d"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.206077 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-gwmvk"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.214428 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-6252-account-create-update-cz7bz"] Jan 06 15:07:56 crc kubenswrapper[4744]: I0106 15:07:56.222774 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-67ffc"] Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.234596 4744 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6fnxs container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.234979 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6fnxs" podUID="e615c362-aa4f-488f-ba46-9fbb9c66e204" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.397614 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16de9a20-7669-4935-9995-a5d1172e68e3" path="/var/lib/kubelet/pods/16de9a20-7669-4935-9995-a5d1172e68e3/volumes" Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.399988 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41cf511d-549f-4b68-bb95-40a73cddbf83" path="/var/lib/kubelet/pods/41cf511d-549f-4b68-bb95-40a73cddbf83/volumes" Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.402001 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b3ba14-81e9-4338-bc3e-297128ea5ef1" path="/var/lib/kubelet/pods/44b3ba14-81e9-4338-bc3e-297128ea5ef1/volumes" Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.404240 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5df8f82c-aba7-4d44-b4bd-c3a616afc480" path="/var/lib/kubelet/pods/5df8f82c-aba7-4d44-b4bd-c3a616afc480/volumes" Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.407057 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="795cf078-881c-4373-9e3c-13afd8662252" path="/var/lib/kubelet/pods/795cf078-881c-4373-9e3c-13afd8662252/volumes" Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.410323 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a372a06b-1453-4141-b82b-96af30ceb608" path="/var/lib/kubelet/pods/a372a06b-1453-4141-b82b-96af30ceb608/volumes" Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.414361 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a57dc900-6506-4a69-b49a-3bd5dd6a88db" path="/var/lib/kubelet/pods/a57dc900-6506-4a69-b49a-3bd5dd6a88db/volumes" Jan 06 15:07:58 crc kubenswrapper[4744]: I0106 15:07:58.416454 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afdb36a9-143a-4bbc-a826-cdfa078e5d68" path="/var/lib/kubelet/pods/afdb36a9-143a-4bbc-a826-cdfa078e5d68/volumes" Jan 06 15:07:59 crc kubenswrapper[4744]: I0106 15:07:59.440353 4744 generic.go:334] "Generic (PLEG): container finished" podID="b4c6a122-8746-4f2b-a7f9-e3df62969170" containerID="aa505f7cef8367ccdaffa0f7a55a505803b66e35822b89aa6f786905ee0543b2" exitCode=0 Jan 06 15:07:59 crc kubenswrapper[4744]: I0106 15:07:59.440443 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b4c6a122-8746-4f2b-a7f9-e3df62969170","Type":"ContainerDied","Data":"aa505f7cef8367ccdaffa0f7a55a505803b66e35822b89aa6f786905ee0543b2"} Jan 06 15:08:06 crc kubenswrapper[4744]: I0106 15:08:06.537023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b4c6a122-8746-4f2b-a7f9-e3df62969170","Type":"ContainerStarted","Data":"0a6a69d35c51f382ac251099885d4b8e86f25791a3a4ee78ab0e90b804d2d97b"} Jan 06 15:08:06 crc kubenswrapper[4744]: I0106 15:08:06.538328 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 06 15:08:06 crc kubenswrapper[4744]: I0106 15:08:06.579824 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.579802948 podStartE2EDuration="50.579802948s" podCreationTimestamp="2026-01-06 15:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:08:06.566360921 +0000 UTC m=+1883.193827249" watchObservedRunningTime="2026-01-06 15:08:06.579802948 +0000 UTC m=+1883.207269266" Jan 06 15:08:08 crc kubenswrapper[4744]: I0106 15:08:08.545969 4744 scope.go:117] "RemoveContainer" containerID="89c738924cdbe81aff4c744ecbdb2598b0a9cb084e30bfe98239d5830b404c4a" Jan 06 15:08:08 crc kubenswrapper[4744]: I0106 15:08:08.599462 4744 scope.go:117] "RemoveContainer" containerID="ceb63e211536bb28cc81a6b2ffd399f52ebccc4932b59313e1a048925e42974d" Jan 06 15:08:08 crc kubenswrapper[4744]: I0106 15:08:08.667931 4744 scope.go:117] "RemoveContainer" containerID="f70b7515706c1872c6ba3959ae5e440e2879d46ac115e47926c935ae30fd139e" Jan 06 15:08:08 crc kubenswrapper[4744]: I0106 15:08:08.733489 4744 scope.go:117] "RemoveContainer" containerID="fa60c28e4cd5fd3128b80c2e4ba583fb162a3418f60a1034dc3f7be5406bd078" Jan 06 15:08:08 crc kubenswrapper[4744]: I0106 15:08:08.794847 4744 scope.go:117] "RemoveContainer" containerID="d5f05e138cfafdf48e62b274ed67756d3c919ceacba303f614f37f053789fcfb" Jan 06 15:08:08 crc kubenswrapper[4744]: I0106 15:08:08.857415 4744 scope.go:117] "RemoveContainer" containerID="897553c86d3651136360bd9d6a4f409073678c53836195c8462b58e9ad930617" Jan 06 15:08:08 crc kubenswrapper[4744]: I0106 15:08:08.893572 4744 scope.go:117] "RemoveContainer" containerID="ba10af281287d6f6d53e1a4f86eb77f6e5a2f7416049570f8be012f584736d87" Jan 06 15:08:08 crc kubenswrapper[4744]: I0106 15:08:08.950590 4744 scope.go:117] "RemoveContainer" containerID="8a8d208339772abdef80cc2136ad05331e6fe7a98777376e84a8b2d48a796568" Jan 06 15:08:08 crc kubenswrapper[4744]: I0106 15:08:08.977095 4744 scope.go:117] "RemoveContainer" containerID="d27c59d832129edb676e4a2f4b297dcfbc8077bc9e58589178053d5fec161fc2" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.004033 4744 scope.go:117] "RemoveContainer" containerID="37171b25c1cefc9d4a58eecf515fcc2ec59aed08d23e508549e1f6bf04fdbd05" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.038611 4744 scope.go:117] "RemoveContainer" containerID="ce294c65bfe21d8d232f8241b1fb908d05a276361b0c60a5be9601443075bf98" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.070630 4744 scope.go:117] "RemoveContainer" containerID="ee305d5dfd8895fcbdc3e6f78124e6aab08b28f9b844ec202b6b0f4730d2722b" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.102926 4744 scope.go:117] "RemoveContainer" containerID="4f27466bb5b9291a0e7e28e0c2cfdf7c509982e49faba5b0ded99d85d4e719fc" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.133854 4744 scope.go:117] "RemoveContainer" containerID="2bf2b634b3024dea2befdb1f467ac97467295a00c1502406a72477235fd0e8a4" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.177878 4744 scope.go:117] "RemoveContainer" containerID="788bcdfaf4046f9a2e160764587f4eb09b4bcbb1d0ea440afaf7d4d09290f34a" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.203959 4744 scope.go:117] "RemoveContainer" containerID="89a23289537705577d0a0e00b5060f8fc126f4bb7c59f24cf2b0282e66862b86" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.237877 4744 scope.go:117] "RemoveContainer" containerID="86c99a2279059f4f70da24573ab4662daad672dc4a447bc5d4cb0b1a275f468b" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.285787 4744 scope.go:117] "RemoveContainer" containerID="4ff9a47398dad8581d91e0b79fe09b122503f679df40098ef216cc80866153c8" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.311467 4744 scope.go:117] "RemoveContainer" containerID="20e35d14624d384244c963aa5a35b1c9117ccb44e1174e5ce9d5eb1e281f03f1" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.352203 4744 scope.go:117] "RemoveContainer" containerID="7b1cba8e587256820a5e1e682c881c55418c80b71d42324099c2c0bafd6134d1" Jan 06 15:08:09 crc kubenswrapper[4744]: I0106 15:08:09.397074 4744 scope.go:117] "RemoveContainer" containerID="10afb9bf4e76dc90e94ac82aeb60ff829b277ed81be73aaa635dc23ca0e5be73" Jan 06 15:08:13 crc kubenswrapper[4744]: I0106 15:08:13.036353 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-jbwmt"] Jan 06 15:08:13 crc kubenswrapper[4744]: I0106 15:08:13.050860 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-jbwmt"] Jan 06 15:08:13 crc kubenswrapper[4744]: I0106 15:08:13.728412 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66f0fe7a-9061-4bc2-a957-f9d2926401d7" path="/var/lib/kubelet/pods/66f0fe7a-9061-4bc2-a957-f9d2926401d7/volumes" Jan 06 15:08:17 crc kubenswrapper[4744]: I0106 15:08:17.594492 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 06 15:08:52 crc kubenswrapper[4744]: I0106 15:08:52.069726 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-68rr4"] Jan 06 15:08:52 crc kubenswrapper[4744]: I0106 15:08:52.081101 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-68rr4"] Jan 06 15:08:53 crc kubenswrapper[4744]: I0106 15:08:53.735745 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb6ad61a-8190-4d4d-987c-f609c1e8cf44" path="/var/lib/kubelet/pods/cb6ad61a-8190-4d4d-987c-f609c1e8cf44/volumes" Jan 06 15:09:09 crc kubenswrapper[4744]: I0106 15:09:09.891076 4744 scope.go:117] "RemoveContainer" containerID="d45a2107bad45f86e44c8fc5363ad14d76ce806baf5a1db686e281efd2dae1c5" Jan 06 15:09:09 crc kubenswrapper[4744]: I0106 15:09:09.947070 4744 scope.go:117] "RemoveContainer" containerID="aa6420ec614c010d558abaa6092c6de2db2940f60b5c9afc7f6a05f7a81c42da" Jan 06 15:09:10 crc kubenswrapper[4744]: I0106 15:09:10.007365 4744 scope.go:117] "RemoveContainer" containerID="ef729b5a2939b5b0e257b95388618106527be7a6dc2ddda746a607596abf9c7e" Jan 06 15:09:10 crc kubenswrapper[4744]: I0106 15:09:10.065601 4744 scope.go:117] "RemoveContainer" containerID="2a70419d4e1df7b45905a583313565b14f3fc14270df6b3cadc52370b7a9d58b" Jan 06 15:09:10 crc kubenswrapper[4744]: I0106 15:09:10.140413 4744 scope.go:117] "RemoveContainer" containerID="2ab20d9a010b2164f31b5440bef346b4bb35d8a31cc6ba96c6bfba16eaf5f6b4" Jan 06 15:09:10 crc kubenswrapper[4744]: I0106 15:09:10.199693 4744 scope.go:117] "RemoveContainer" containerID="e3637667911620da1180a235994b6954085165329abea4bb7a1f23fb84667847" Jan 06 15:09:10 crc kubenswrapper[4744]: I0106 15:09:10.242709 4744 scope.go:117] "RemoveContainer" containerID="bfbda32710913e7ee8d3e6ce097f00ee3bb669ab8a458bb98f80315331a8144e" Jan 06 15:09:12 crc kubenswrapper[4744]: I0106 15:09:12.054200 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-tffgx"] Jan 06 15:09:12 crc kubenswrapper[4744]: I0106 15:09:12.064660 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-tffgx"] Jan 06 15:09:13 crc kubenswrapper[4744]: I0106 15:09:13.723590 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c5385b7-40c0-4927-87fe-3db7adf6cdab" path="/var/lib/kubelet/pods/8c5385b7-40c0-4927-87fe-3db7adf6cdab/volumes" Jan 06 15:09:14 crc kubenswrapper[4744]: I0106 15:09:14.424055 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:09:14 crc kubenswrapper[4744]: I0106 15:09:14.424626 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:09:26 crc kubenswrapper[4744]: I0106 15:09:26.052854 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-dfxz9"] Jan 06 15:09:26 crc kubenswrapper[4744]: I0106 15:09:26.066322 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-dfxz9"] Jan 06 15:09:27 crc kubenswrapper[4744]: I0106 15:09:27.733432 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be9d6eb2-8ad6-4eb4-9208-a89adf751c4b" path="/var/lib/kubelet/pods/be9d6eb2-8ad6-4eb4-9208-a89adf751c4b/volumes" Jan 06 15:09:43 crc kubenswrapper[4744]: I0106 15:09:43.827824 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7hm5f"] Jan 06 15:09:43 crc kubenswrapper[4744]: I0106 15:09:43.848889 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7hm5f"] Jan 06 15:09:43 crc kubenswrapper[4744]: I0106 15:09:43.849112 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:43 crc kubenswrapper[4744]: I0106 15:09:43.976663 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6phcz\" (UniqueName: \"kubernetes.io/projected/00d16d55-49f8-48f5-b008-2cb93a1332a5-kube-api-access-6phcz\") pod \"certified-operators-7hm5f\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:43 crc kubenswrapper[4744]: I0106 15:09:43.976767 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-utilities\") pod \"certified-operators-7hm5f\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:43 crc kubenswrapper[4744]: I0106 15:09:43.976921 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-catalog-content\") pod \"certified-operators-7hm5f\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.079234 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6phcz\" (UniqueName: \"kubernetes.io/projected/00d16d55-49f8-48f5-b008-2cb93a1332a5-kube-api-access-6phcz\") pod \"certified-operators-7hm5f\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.079315 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-utilities\") pod \"certified-operators-7hm5f\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.079412 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-catalog-content\") pod \"certified-operators-7hm5f\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.079951 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-catalog-content\") pod \"certified-operators-7hm5f\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.080243 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-utilities\") pod \"certified-operators-7hm5f\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.100915 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6phcz\" (UniqueName: \"kubernetes.io/projected/00d16d55-49f8-48f5-b008-2cb93a1332a5-kube-api-access-6phcz\") pod \"certified-operators-7hm5f\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.184145 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.451296 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.451533 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:09:44 crc kubenswrapper[4744]: W0106 15:09:44.780126 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00d16d55_49f8_48f5_b008_2cb93a1332a5.slice/crio-0f090ead52349f7165f822a9e54b17ee0c7614ffe8385a8e08a4c8a7f883b543 WatchSource:0}: Error finding container 0f090ead52349f7165f822a9e54b17ee0c7614ffe8385a8e08a4c8a7f883b543: Status 404 returned error can't find the container with id 0f090ead52349f7165f822a9e54b17ee0c7614ffe8385a8e08a4c8a7f883b543 Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.780413 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7hm5f"] Jan 06 15:09:44 crc kubenswrapper[4744]: I0106 15:09:44.805466 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hm5f" event={"ID":"00d16d55-49f8-48f5-b008-2cb93a1332a5","Type":"ContainerStarted","Data":"0f090ead52349f7165f822a9e54b17ee0c7614ffe8385a8e08a4c8a7f883b543"} Jan 06 15:09:45 crc kubenswrapper[4744]: I0106 15:09:45.821114 4744 generic.go:334] "Generic (PLEG): container finished" podID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerID="b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c" exitCode=0 Jan 06 15:09:45 crc kubenswrapper[4744]: I0106 15:09:45.821279 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hm5f" event={"ID":"00d16d55-49f8-48f5-b008-2cb93a1332a5","Type":"ContainerDied","Data":"b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c"} Jan 06 15:09:45 crc kubenswrapper[4744]: I0106 15:09:45.823891 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.185075 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l9lgw"] Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.223342 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9lgw"] Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.223446 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.346015 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4djn\" (UniqueName: \"kubernetes.io/projected/278473b1-b2fe-4dba-857e-ca6c9859169c-kube-api-access-j4djn\") pod \"redhat-operators-l9lgw\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.346092 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-utilities\") pod \"redhat-operators-l9lgw\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.346177 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-catalog-content\") pod \"redhat-operators-l9lgw\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.448380 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-utilities\") pod \"redhat-operators-l9lgw\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.448615 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-catalog-content\") pod \"redhat-operators-l9lgw\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.448990 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-utilities\") pod \"redhat-operators-l9lgw\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.449119 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-catalog-content\") pod \"redhat-operators-l9lgw\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.449122 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4djn\" (UniqueName: \"kubernetes.io/projected/278473b1-b2fe-4dba-857e-ca6c9859169c-kube-api-access-j4djn\") pod \"redhat-operators-l9lgw\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.478196 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4djn\" (UniqueName: \"kubernetes.io/projected/278473b1-b2fe-4dba-857e-ca6c9859169c-kube-api-access-j4djn\") pod \"redhat-operators-l9lgw\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:46 crc kubenswrapper[4744]: I0106 15:09:46.543949 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:09:47 crc kubenswrapper[4744]: I0106 15:09:47.078129 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9lgw"] Jan 06 15:09:47 crc kubenswrapper[4744]: I0106 15:09:47.865667 4744 generic.go:334] "Generic (PLEG): container finished" podID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerID="a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7" exitCode=0 Jan 06 15:09:47 crc kubenswrapper[4744]: I0106 15:09:47.865873 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lgw" event={"ID":"278473b1-b2fe-4dba-857e-ca6c9859169c","Type":"ContainerDied","Data":"a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7"} Jan 06 15:09:47 crc kubenswrapper[4744]: I0106 15:09:47.866282 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lgw" event={"ID":"278473b1-b2fe-4dba-857e-ca6c9859169c","Type":"ContainerStarted","Data":"80c10a26745779bc02735de7d1da124c16aaa2d6fa196134963518ff17db71fe"} Jan 06 15:09:47 crc kubenswrapper[4744]: I0106 15:09:47.878851 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hm5f" event={"ID":"00d16d55-49f8-48f5-b008-2cb93a1332a5","Type":"ContainerStarted","Data":"61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee"} Jan 06 15:09:52 crc kubenswrapper[4744]: I0106 15:09:52.932232 4744 generic.go:334] "Generic (PLEG): container finished" podID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerID="61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee" exitCode=0 Jan 06 15:09:52 crc kubenswrapper[4744]: I0106 15:09:52.932290 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hm5f" event={"ID":"00d16d55-49f8-48f5-b008-2cb93a1332a5","Type":"ContainerDied","Data":"61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee"} Jan 06 15:09:52 crc kubenswrapper[4744]: I0106 15:09:52.935723 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lgw" event={"ID":"278473b1-b2fe-4dba-857e-ca6c9859169c","Type":"ContainerStarted","Data":"b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a"} Jan 06 15:09:54 crc kubenswrapper[4744]: I0106 15:09:54.970892 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hm5f" event={"ID":"00d16d55-49f8-48f5-b008-2cb93a1332a5","Type":"ContainerStarted","Data":"b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c"} Jan 06 15:09:55 crc kubenswrapper[4744]: I0106 15:09:55.007604 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7hm5f" podStartSLOduration=4.074304689 podStartE2EDuration="12.007578156s" podCreationTimestamp="2026-01-06 15:09:43 +0000 UTC" firstStartedPulling="2026-01-06 15:09:45.823651189 +0000 UTC m=+1982.451117507" lastFinishedPulling="2026-01-06 15:09:53.756924656 +0000 UTC m=+1990.384390974" observedRunningTime="2026-01-06 15:09:55.004637698 +0000 UTC m=+1991.632104056" watchObservedRunningTime="2026-01-06 15:09:55.007578156 +0000 UTC m=+1991.635044524" Jan 06 15:09:59 crc kubenswrapper[4744]: I0106 15:09:59.048831 4744 generic.go:334] "Generic (PLEG): container finished" podID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerID="b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a" exitCode=0 Jan 06 15:09:59 crc kubenswrapper[4744]: I0106 15:09:59.048878 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lgw" event={"ID":"278473b1-b2fe-4dba-857e-ca6c9859169c","Type":"ContainerDied","Data":"b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a"} Jan 06 15:10:00 crc kubenswrapper[4744]: I0106 15:10:00.076210 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lgw" event={"ID":"278473b1-b2fe-4dba-857e-ca6c9859169c","Type":"ContainerStarted","Data":"05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515"} Jan 06 15:10:00 crc kubenswrapper[4744]: I0106 15:10:00.113151 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l9lgw" podStartSLOduration=2.491529435 podStartE2EDuration="14.113125513s" podCreationTimestamp="2026-01-06 15:09:46 +0000 UTC" firstStartedPulling="2026-01-06 15:09:47.872309126 +0000 UTC m=+1984.499775454" lastFinishedPulling="2026-01-06 15:09:59.493905204 +0000 UTC m=+1996.121371532" observedRunningTime="2026-01-06 15:10:00.107106713 +0000 UTC m=+1996.734573081" watchObservedRunningTime="2026-01-06 15:10:00.113125513 +0000 UTC m=+1996.740591831" Jan 06 15:10:04 crc kubenswrapper[4744]: I0106 15:10:04.185062 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:10:04 crc kubenswrapper[4744]: I0106 15:10:04.185570 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:10:04 crc kubenswrapper[4744]: I0106 15:10:04.235135 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:10:05 crc kubenswrapper[4744]: I0106 15:10:05.043140 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-m8c54"] Jan 06 15:10:05 crc kubenswrapper[4744]: I0106 15:10:05.064524 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-m8c54"] Jan 06 15:10:05 crc kubenswrapper[4744]: I0106 15:10:05.216110 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:10:05 crc kubenswrapper[4744]: I0106 15:10:05.266519 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7hm5f"] Jan 06 15:10:05 crc kubenswrapper[4744]: I0106 15:10:05.724994 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d50465b-09ea-4ff4-aa47-12f1c49ff207" path="/var/lib/kubelet/pods/4d50465b-09ea-4ff4-aa47-12f1c49ff207/volumes" Jan 06 15:10:06 crc kubenswrapper[4744]: I0106 15:10:06.544087 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:10:06 crc kubenswrapper[4744]: I0106 15:10:06.544512 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:10:07 crc kubenswrapper[4744]: I0106 15:10:07.182378 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7hm5f" podUID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerName="registry-server" containerID="cri-o://b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c" gracePeriod=2 Jan 06 15:10:07 crc kubenswrapper[4744]: I0106 15:10:07.634537 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9lgw" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerName="registry-server" probeResult="failure" output=< Jan 06 15:10:07 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 15:10:07 crc kubenswrapper[4744]: > Jan 06 15:10:07 crc kubenswrapper[4744]: I0106 15:10:07.822327 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:10:07 crc kubenswrapper[4744]: I0106 15:10:07.918908 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6phcz\" (UniqueName: \"kubernetes.io/projected/00d16d55-49f8-48f5-b008-2cb93a1332a5-kube-api-access-6phcz\") pod \"00d16d55-49f8-48f5-b008-2cb93a1332a5\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " Jan 06 15:10:07 crc kubenswrapper[4744]: I0106 15:10:07.919225 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-catalog-content\") pod \"00d16d55-49f8-48f5-b008-2cb93a1332a5\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " Jan 06 15:10:07 crc kubenswrapper[4744]: I0106 15:10:07.919392 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-utilities\") pod \"00d16d55-49f8-48f5-b008-2cb93a1332a5\" (UID: \"00d16d55-49f8-48f5-b008-2cb93a1332a5\") " Jan 06 15:10:07 crc kubenswrapper[4744]: I0106 15:10:07.920299 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-utilities" (OuterVolumeSpecName: "utilities") pod "00d16d55-49f8-48f5-b008-2cb93a1332a5" (UID: "00d16d55-49f8-48f5-b008-2cb93a1332a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:10:07 crc kubenswrapper[4744]: I0106 15:10:07.928220 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d16d55-49f8-48f5-b008-2cb93a1332a5-kube-api-access-6phcz" (OuterVolumeSpecName: "kube-api-access-6phcz") pod "00d16d55-49f8-48f5-b008-2cb93a1332a5" (UID: "00d16d55-49f8-48f5-b008-2cb93a1332a5"). InnerVolumeSpecName "kube-api-access-6phcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:10:07 crc kubenswrapper[4744]: I0106 15:10:07.973792 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "00d16d55-49f8-48f5-b008-2cb93a1332a5" (UID: "00d16d55-49f8-48f5-b008-2cb93a1332a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.021814 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.022118 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/00d16d55-49f8-48f5-b008-2cb93a1332a5-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.022129 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6phcz\" (UniqueName: \"kubernetes.io/projected/00d16d55-49f8-48f5-b008-2cb93a1332a5-kube-api-access-6phcz\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.195084 4744 generic.go:334] "Generic (PLEG): container finished" podID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerID="b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c" exitCode=0 Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.195192 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7hm5f" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.195185 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hm5f" event={"ID":"00d16d55-49f8-48f5-b008-2cb93a1332a5","Type":"ContainerDied","Data":"b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c"} Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.195271 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7hm5f" event={"ID":"00d16d55-49f8-48f5-b008-2cb93a1332a5","Type":"ContainerDied","Data":"0f090ead52349f7165f822a9e54b17ee0c7614ffe8385a8e08a4c8a7f883b543"} Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.195308 4744 scope.go:117] "RemoveContainer" containerID="b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.230635 4744 scope.go:117] "RemoveContainer" containerID="61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.247758 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7hm5f"] Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.259498 4744 scope.go:117] "RemoveContainer" containerID="b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.261477 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7hm5f"] Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.338758 4744 scope.go:117] "RemoveContainer" containerID="b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c" Jan 06 15:10:08 crc kubenswrapper[4744]: E0106 15:10:08.340655 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c\": container with ID starting with b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c not found: ID does not exist" containerID="b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.340713 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c"} err="failed to get container status \"b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c\": rpc error: code = NotFound desc = could not find container \"b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c\": container with ID starting with b8ea7a375f0a1d5ff6ef4a2935417f5c792579ec38dd64118612e72e692bce3c not found: ID does not exist" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.340751 4744 scope.go:117] "RemoveContainer" containerID="61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee" Jan 06 15:10:08 crc kubenswrapper[4744]: E0106 15:10:08.341188 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee\": container with ID starting with 61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee not found: ID does not exist" containerID="61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.341262 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee"} err="failed to get container status \"61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee\": rpc error: code = NotFound desc = could not find container \"61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee\": container with ID starting with 61d6ba3564bc9f09241555c0febf76c63aaa72bd5a486dab4db014fd992981ee not found: ID does not exist" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.341317 4744 scope.go:117] "RemoveContainer" containerID="b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c" Jan 06 15:10:08 crc kubenswrapper[4744]: E0106 15:10:08.341695 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c\": container with ID starting with b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c not found: ID does not exist" containerID="b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c" Jan 06 15:10:08 crc kubenswrapper[4744]: I0106 15:10:08.341748 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c"} err="failed to get container status \"b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c\": rpc error: code = NotFound desc = could not find container \"b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c\": container with ID starting with b316d65345a6d6984b30d6d2678d95f56edd013ce3e657265671117a4557589c not found: ID does not exist" Jan 06 15:10:09 crc kubenswrapper[4744]: I0106 15:10:09.725881 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00d16d55-49f8-48f5-b008-2cb93a1332a5" path="/var/lib/kubelet/pods/00d16d55-49f8-48f5-b008-2cb93a1332a5/volumes" Jan 06 15:10:10 crc kubenswrapper[4744]: I0106 15:10:10.390198 4744 scope.go:117] "RemoveContainer" containerID="ff7d2ade3aab7c9781162b8f5dbd880658f6b48bae8d4bef5e1ff547a738b44f" Jan 06 15:10:10 crc kubenswrapper[4744]: I0106 15:10:10.461381 4744 scope.go:117] "RemoveContainer" containerID="c948c1d0335191aa70e62006bf06381bf3b03ffdaa262d1ef0601687003ff060" Jan 06 15:10:10 crc kubenswrapper[4744]: I0106 15:10:10.507362 4744 scope.go:117] "RemoveContainer" containerID="6dbf2f5d4c23d880674670d0d4a0482afb770c9fd6ec883381a24e1ed079b285" Jan 06 15:10:13 crc kubenswrapper[4744]: I0106 15:10:13.035317 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-m88z9"] Jan 06 15:10:13 crc kubenswrapper[4744]: I0106 15:10:13.047636 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-m88z9"] Jan 06 15:10:13 crc kubenswrapper[4744]: I0106 15:10:13.735958 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9e793d1-65f3-43a8-914b-112690afc33f" path="/var/lib/kubelet/pods/d9e793d1-65f3-43a8-914b-112690afc33f/volumes" Jan 06 15:10:14 crc kubenswrapper[4744]: I0106 15:10:14.423851 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:10:14 crc kubenswrapper[4744]: I0106 15:10:14.423914 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:10:14 crc kubenswrapper[4744]: I0106 15:10:14.423963 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:10:14 crc kubenswrapper[4744]: I0106 15:10:14.425022 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d70e5126160c45c7349d68db9fd09dd06c15ea7a0105b790d710f8a5423e632"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:10:14 crc kubenswrapper[4744]: I0106 15:10:14.425112 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://8d70e5126160c45c7349d68db9fd09dd06c15ea7a0105b790d710f8a5423e632" gracePeriod=600 Jan 06 15:10:15 crc kubenswrapper[4744]: I0106 15:10:15.042528 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-vdnt5"] Jan 06 15:10:15 crc kubenswrapper[4744]: I0106 15:10:15.060615 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-vdnt5"] Jan 06 15:10:15 crc kubenswrapper[4744]: I0106 15:10:15.303137 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="8d70e5126160c45c7349d68db9fd09dd06c15ea7a0105b790d710f8a5423e632" exitCode=0 Jan 06 15:10:15 crc kubenswrapper[4744]: I0106 15:10:15.303182 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"8d70e5126160c45c7349d68db9fd09dd06c15ea7a0105b790d710f8a5423e632"} Jan 06 15:10:15 crc kubenswrapper[4744]: I0106 15:10:15.303220 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce"} Jan 06 15:10:15 crc kubenswrapper[4744]: I0106 15:10:15.303237 4744 scope.go:117] "RemoveContainer" containerID="2dd1ff7bd038d4d91b2a61b8f62f55bed45600531104c1cdf4f3034dce85d61c" Jan 06 15:10:15 crc kubenswrapper[4744]: I0106 15:10:15.724356 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b74efd3-9c21-4555-bfa8-5c1652b380dc" path="/var/lib/kubelet/pods/2b74efd3-9c21-4555-bfa8-5c1652b380dc/volumes" Jan 06 15:10:16 crc kubenswrapper[4744]: I0106 15:10:16.632702 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:10:16 crc kubenswrapper[4744]: I0106 15:10:16.687896 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:10:17 crc kubenswrapper[4744]: I0106 15:10:17.387882 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9lgw"] Jan 06 15:10:18 crc kubenswrapper[4744]: I0106 15:10:18.346230 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l9lgw" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerName="registry-server" containerID="cri-o://05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515" gracePeriod=2 Jan 06 15:10:18 crc kubenswrapper[4744]: E0106 15:10:18.687057 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod278473b1_b2fe_4dba_857e_ca6c9859169c.slice/crio-05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod278473b1_b2fe_4dba_857e_ca6c9859169c.slice/crio-conmon-05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:10:18 crc kubenswrapper[4744]: I0106 15:10:18.882892 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:10:18 crc kubenswrapper[4744]: I0106 15:10:18.999873 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-catalog-content\") pod \"278473b1-b2fe-4dba-857e-ca6c9859169c\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:18.999951 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4djn\" (UniqueName: \"kubernetes.io/projected/278473b1-b2fe-4dba-857e-ca6c9859169c-kube-api-access-j4djn\") pod \"278473b1-b2fe-4dba-857e-ca6c9859169c\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.000132 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-utilities\") pod \"278473b1-b2fe-4dba-857e-ca6c9859169c\" (UID: \"278473b1-b2fe-4dba-857e-ca6c9859169c\") " Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.000912 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-utilities" (OuterVolumeSpecName: "utilities") pod "278473b1-b2fe-4dba-857e-ca6c9859169c" (UID: "278473b1-b2fe-4dba-857e-ca6c9859169c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.015400 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/278473b1-b2fe-4dba-857e-ca6c9859169c-kube-api-access-j4djn" (OuterVolumeSpecName: "kube-api-access-j4djn") pod "278473b1-b2fe-4dba-857e-ca6c9859169c" (UID: "278473b1-b2fe-4dba-857e-ca6c9859169c"). InnerVolumeSpecName "kube-api-access-j4djn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.109916 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.110415 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4djn\" (UniqueName: \"kubernetes.io/projected/278473b1-b2fe-4dba-857e-ca6c9859169c-kube-api-access-j4djn\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.148996 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "278473b1-b2fe-4dba-857e-ca6c9859169c" (UID: "278473b1-b2fe-4dba-857e-ca6c9859169c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.213083 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/278473b1-b2fe-4dba-857e-ca6c9859169c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.357987 4744 generic.go:334] "Generic (PLEG): container finished" podID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerID="05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515" exitCode=0 Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.358059 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9lgw" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.358059 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lgw" event={"ID":"278473b1-b2fe-4dba-857e-ca6c9859169c","Type":"ContainerDied","Data":"05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515"} Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.363282 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9lgw" event={"ID":"278473b1-b2fe-4dba-857e-ca6c9859169c","Type":"ContainerDied","Data":"80c10a26745779bc02735de7d1da124c16aaa2d6fa196134963518ff17db71fe"} Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.363318 4744 scope.go:117] "RemoveContainer" containerID="05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.390530 4744 scope.go:117] "RemoveContainer" containerID="b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.410238 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9lgw"] Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.423166 4744 scope.go:117] "RemoveContainer" containerID="a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.448786 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l9lgw"] Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.487653 4744 scope.go:117] "RemoveContainer" containerID="05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515" Jan 06 15:10:19 crc kubenswrapper[4744]: E0106 15:10:19.488436 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515\": container with ID starting with 05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515 not found: ID does not exist" containerID="05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.488685 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515"} err="failed to get container status \"05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515\": rpc error: code = NotFound desc = could not find container \"05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515\": container with ID starting with 05b3c6f1ca27c3b2dc36c570f064ca077e7afc1cf8134347cb21529f9abc1515 not found: ID does not exist" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.488715 4744 scope.go:117] "RemoveContainer" containerID="b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a" Jan 06 15:10:19 crc kubenswrapper[4744]: E0106 15:10:19.489196 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a\": container with ID starting with b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a not found: ID does not exist" containerID="b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.489246 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a"} err="failed to get container status \"b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a\": rpc error: code = NotFound desc = could not find container \"b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a\": container with ID starting with b5a603467bb15548f856059f8777074fc99c628e5a0b25667df29ac4132c906a not found: ID does not exist" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.489280 4744 scope.go:117] "RemoveContainer" containerID="a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7" Jan 06 15:10:19 crc kubenswrapper[4744]: E0106 15:10:19.489795 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7\": container with ID starting with a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7 not found: ID does not exist" containerID="a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.489847 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7"} err="failed to get container status \"a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7\": rpc error: code = NotFound desc = could not find container \"a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7\": container with ID starting with a52d08a0bdf480a70f4258e9ef44b8bb4358c663db750c462712ce29f5398ca7 not found: ID does not exist" Jan 06 15:10:19 crc kubenswrapper[4744]: I0106 15:10:19.732355 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" path="/var/lib/kubelet/pods/278473b1-b2fe-4dba-857e-ca6c9859169c/volumes" Jan 06 15:10:37 crc kubenswrapper[4744]: I0106 15:10:37.567298 4744 generic.go:334] "Generic (PLEG): container finished" podID="eac03a71-e859-4391-b58e-f4dc4fdfa0ce" containerID="d932e43f1febe26a75c11afe102a8d6bd107eca7366410e9cecb37db8f233871" exitCode=0 Jan 06 15:10:37 crc kubenswrapper[4744]: I0106 15:10:37.567445 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" event={"ID":"eac03a71-e859-4391-b58e-f4dc4fdfa0ce","Type":"ContainerDied","Data":"d932e43f1febe26a75c11afe102a8d6bd107eca7366410e9cecb37db8f233871"} Jan 06 15:10:38 crc kubenswrapper[4744]: I0106 15:10:38.053636 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jvkwj"] Jan 06 15:10:38 crc kubenswrapper[4744]: I0106 15:10:38.071250 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jvkwj"] Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.040394 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-afb3-account-create-update-k5htr"] Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.065049 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-sk49l"] Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.075208 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-b7ad-account-create-update-s4ph8"] Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.085828 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-afb3-account-create-update-k5htr"] Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.095442 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-sk49l"] Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.107498 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-b7ad-account-create-update-s4ph8"] Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.158919 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.181330 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-ssh-key-openstack-edpm-ipam\") pod \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.181717 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-inventory\") pod \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.216415 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "eac03a71-e859-4391-b58e-f4dc4fdfa0ce" (UID: "eac03a71-e859-4391-b58e-f4dc4fdfa0ce"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.230388 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-inventory" (OuterVolumeSpecName: "inventory") pod "eac03a71-e859-4391-b58e-f4dc4fdfa0ce" (UID: "eac03a71-e859-4391-b58e-f4dc4fdfa0ce"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.283545 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-bootstrap-combined-ca-bundle\") pod \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.283894 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zngx9\" (UniqueName: \"kubernetes.io/projected/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-kube-api-access-zngx9\") pod \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\" (UID: \"eac03a71-e859-4391-b58e-f4dc4fdfa0ce\") " Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.285904 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.285932 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.294582 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-kube-api-access-zngx9" (OuterVolumeSpecName: "kube-api-access-zngx9") pod "eac03a71-e859-4391-b58e-f4dc4fdfa0ce" (UID: "eac03a71-e859-4391-b58e-f4dc4fdfa0ce"). InnerVolumeSpecName "kube-api-access-zngx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.294764 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "eac03a71-e859-4391-b58e-f4dc4fdfa0ce" (UID: "eac03a71-e859-4391-b58e-f4dc4fdfa0ce"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.387551 4744 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.387809 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zngx9\" (UniqueName: \"kubernetes.io/projected/eac03a71-e859-4391-b58e-f4dc4fdfa0ce-kube-api-access-zngx9\") on node \"crc\" DevicePath \"\"" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.593029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" event={"ID":"eac03a71-e859-4391-b58e-f4dc4fdfa0ce","Type":"ContainerDied","Data":"b5496902710052f46c0aa76fd721cdd07542a479ee5e3e740144b650244b0a6f"} Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.593074 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5496902710052f46c0aa76fd721cdd07542a479ee5e3e740144b650244b0a6f" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.593142 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.699768 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5"] Jan 06 15:10:39 crc kubenswrapper[4744]: E0106 15:10:39.700456 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerName="extract-content" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700472 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerName="extract-content" Jan 06 15:10:39 crc kubenswrapper[4744]: E0106 15:10:39.700487 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerName="extract-content" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700493 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerName="extract-content" Jan 06 15:10:39 crc kubenswrapper[4744]: E0106 15:10:39.700506 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerName="registry-server" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700514 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerName="registry-server" Jan 06 15:10:39 crc kubenswrapper[4744]: E0106 15:10:39.700520 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerName="registry-server" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700525 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerName="registry-server" Jan 06 15:10:39 crc kubenswrapper[4744]: E0106 15:10:39.700543 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eac03a71-e859-4391-b58e-f4dc4fdfa0ce" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700549 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="eac03a71-e859-4391-b58e-f4dc4fdfa0ce" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 06 15:10:39 crc kubenswrapper[4744]: E0106 15:10:39.700580 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerName="extract-utilities" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700591 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerName="extract-utilities" Jan 06 15:10:39 crc kubenswrapper[4744]: E0106 15:10:39.700610 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerName="extract-utilities" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700615 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerName="extract-utilities" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700801 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d16d55-49f8-48f5-b008-2cb93a1332a5" containerName="registry-server" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700815 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="eac03a71-e859-4391-b58e-f4dc4fdfa0ce" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.700829 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="278473b1-b2fe-4dba-857e-ca6c9859169c" containerName="registry-server" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.701660 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.704824 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.705101 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.705917 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.706363 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.728278 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1174ed04-5464-450f-850f-e77c11afee5d" path="/var/lib/kubelet/pods/1174ed04-5464-450f-850f-e77c11afee5d/volumes" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.729485 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c698d4-2488-47f1-ad67-7506ec912b38" path="/var/lib/kubelet/pods/49c698d4-2488-47f1-ad67-7506ec912b38/volumes" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.730490 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2afd989-96a7-4562-9ffd-1fac6a8f3acd" path="/var/lib/kubelet/pods/c2afd989-96a7-4562-9ffd-1fac6a8f3acd/volumes" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.731487 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e57295aa-50a7-478a-805c-a682bcf0386b" path="/var/lib/kubelet/pods/e57295aa-50a7-478a-805c-a682bcf0386b/volumes" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.734484 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5"] Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.795353 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.795418 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.795566 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvpzs\" (UniqueName: \"kubernetes.io/projected/98606045-db2f-49f7-a0a3-c88b391364c6-kube-api-access-mvpzs\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.898529 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvpzs\" (UniqueName: \"kubernetes.io/projected/98606045-db2f-49f7-a0a3-c88b391364c6-kube-api-access-mvpzs\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.898711 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.898767 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.902487 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.903813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:39 crc kubenswrapper[4744]: I0106 15:10:39.915068 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvpzs\" (UniqueName: \"kubernetes.io/projected/98606045-db2f-49f7-a0a3-c88b391364c6-kube-api-access-mvpzs\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:40 crc kubenswrapper[4744]: I0106 15:10:40.032699 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:10:40 crc kubenswrapper[4744]: I0106 15:10:40.581034 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5"] Jan 06 15:10:40 crc kubenswrapper[4744]: I0106 15:10:40.604223 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" event={"ID":"98606045-db2f-49f7-a0a3-c88b391364c6","Type":"ContainerStarted","Data":"45c421d0e1e5cdd9358fa497a6db4f71dcd7981850eefdb994b618101bfaedd4"} Jan 06 15:10:41 crc kubenswrapper[4744]: I0106 15:10:41.037083 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-8jcdj"] Jan 06 15:10:41 crc kubenswrapper[4744]: I0106 15:10:41.055528 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-63fc-account-create-update-bjlhb"] Jan 06 15:10:41 crc kubenswrapper[4744]: I0106 15:10:41.067062 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-8jcdj"] Jan 06 15:10:41 crc kubenswrapper[4744]: I0106 15:10:41.078874 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-63fc-account-create-update-bjlhb"] Jan 06 15:10:41 crc kubenswrapper[4744]: I0106 15:10:41.629402 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" event={"ID":"98606045-db2f-49f7-a0a3-c88b391364c6","Type":"ContainerStarted","Data":"cbfe473dca0173764cd2dd40d74844ed84c7312db4282d71da0c653dc8f19894"} Jan 06 15:10:41 crc kubenswrapper[4744]: I0106 15:10:41.663684 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" podStartSLOduration=2.0459002 podStartE2EDuration="2.663662368s" podCreationTimestamp="2026-01-06 15:10:39 +0000 UTC" firstStartedPulling="2026-01-06 15:10:40.591934671 +0000 UTC m=+2037.219400989" lastFinishedPulling="2026-01-06 15:10:41.209696839 +0000 UTC m=+2037.837163157" observedRunningTime="2026-01-06 15:10:41.651423453 +0000 UTC m=+2038.278889771" watchObservedRunningTime="2026-01-06 15:10:41.663662368 +0000 UTC m=+2038.291128696" Jan 06 15:10:41 crc kubenswrapper[4744]: I0106 15:10:41.728964 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49796ee8-5c97-4c55-a8d8-4c47279d2641" path="/var/lib/kubelet/pods/49796ee8-5c97-4c55-a8d8-4c47279d2641/volumes" Jan 06 15:10:41 crc kubenswrapper[4744]: I0106 15:10:41.730443 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d1c4236-2610-4bcd-91f7-c157f5da7ea4" path="/var/lib/kubelet/pods/6d1c4236-2610-4bcd-91f7-c157f5da7ea4/volumes" Jan 06 15:11:10 crc kubenswrapper[4744]: I0106 15:11:10.639430 4744 scope.go:117] "RemoveContainer" containerID="a4deba525591147b2da1e77d8a28191c995aff4a99f4ef3dd3ef23257f19c60a" Jan 06 15:11:10 crc kubenswrapper[4744]: I0106 15:11:10.690762 4744 scope.go:117] "RemoveContainer" containerID="70a155e38d038bc8d794f5c4229b8462818b4baae2a1afd8640b477bdf7710e3" Jan 06 15:11:10 crc kubenswrapper[4744]: I0106 15:11:10.738862 4744 scope.go:117] "RemoveContainer" containerID="7fc6cf3ccd66f28305831ae48e4b4a1ff31ea27fb84df4dd8b0a69e938aec37b" Jan 06 15:11:10 crc kubenswrapper[4744]: I0106 15:11:10.790634 4744 scope.go:117] "RemoveContainer" containerID="032bc9033f646fb0f2a23fc2d0462ef4c3fdb4b022205c7168d5783f751191c4" Jan 06 15:11:10 crc kubenswrapper[4744]: I0106 15:11:10.845260 4744 scope.go:117] "RemoveContainer" containerID="2f2600111ba19818cd3e8825d0359df3c02a05a77cdbe33bd933478b8cd73829" Jan 06 15:11:10 crc kubenswrapper[4744]: I0106 15:11:10.905425 4744 scope.go:117] "RemoveContainer" containerID="b7bdb3b47955e4488022736ca45ed699e39c0136c090450314114a05683ad88e" Jan 06 15:11:10 crc kubenswrapper[4744]: I0106 15:11:10.955395 4744 scope.go:117] "RemoveContainer" containerID="d2ba9c57984ae211313cabc1cac7e28a4fc4db714de0a470aaa2cc18bf0e8792" Jan 06 15:11:10 crc kubenswrapper[4744]: I0106 15:11:10.982799 4744 scope.go:117] "RemoveContainer" containerID="814e34f850c30260a58a4f62d13a45bad787858a2a98db97339a39daf5eb963d" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.145304 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jxvg5"] Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.149581 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.160485 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-catalog-content\") pod \"community-operators-jxvg5\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.160553 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv8fz\" (UniqueName: \"kubernetes.io/projected/47934586-5fc0-47bf-bae3-e7dcd356e98c-kube-api-access-bv8fz\") pod \"community-operators-jxvg5\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.160578 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-utilities\") pod \"community-operators-jxvg5\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.195245 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jxvg5"] Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.269611 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-catalog-content\") pod \"community-operators-jxvg5\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.270418 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv8fz\" (UniqueName: \"kubernetes.io/projected/47934586-5fc0-47bf-bae3-e7dcd356e98c-kube-api-access-bv8fz\") pod \"community-operators-jxvg5\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.270950 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-utilities\") pod \"community-operators-jxvg5\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.271918 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-utilities\") pod \"community-operators-jxvg5\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.272282 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-catalog-content\") pod \"community-operators-jxvg5\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.338032 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv8fz\" (UniqueName: \"kubernetes.io/projected/47934586-5fc0-47bf-bae3-e7dcd356e98c-kube-api-access-bv8fz\") pod \"community-operators-jxvg5\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:19 crc kubenswrapper[4744]: I0106 15:11:19.487141 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:20 crc kubenswrapper[4744]: I0106 15:11:20.016602 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jxvg5"] Jan 06 15:11:20 crc kubenswrapper[4744]: I0106 15:11:20.122066 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxvg5" event={"ID":"47934586-5fc0-47bf-bae3-e7dcd356e98c","Type":"ContainerStarted","Data":"8a1c68dbd104f5b045b05b47b09da9e354ca985822af4e8851f5fdc952e5888f"} Jan 06 15:11:21 crc kubenswrapper[4744]: I0106 15:11:21.141317 4744 generic.go:334] "Generic (PLEG): container finished" podID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerID="8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37" exitCode=0 Jan 06 15:11:21 crc kubenswrapper[4744]: I0106 15:11:21.141409 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxvg5" event={"ID":"47934586-5fc0-47bf-bae3-e7dcd356e98c","Type":"ContainerDied","Data":"8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37"} Jan 06 15:11:23 crc kubenswrapper[4744]: I0106 15:11:23.171536 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxvg5" event={"ID":"47934586-5fc0-47bf-bae3-e7dcd356e98c","Type":"ContainerStarted","Data":"ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa"} Jan 06 15:11:24 crc kubenswrapper[4744]: E0106 15:11:24.059835 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47934586_5fc0_47bf_bae3_e7dcd356e98c.slice/crio-ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:11:25 crc kubenswrapper[4744]: I0106 15:11:25.199504 4744 generic.go:334] "Generic (PLEG): container finished" podID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerID="ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa" exitCode=0 Jan 06 15:11:25 crc kubenswrapper[4744]: I0106 15:11:25.199568 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxvg5" event={"ID":"47934586-5fc0-47bf-bae3-e7dcd356e98c","Type":"ContainerDied","Data":"ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa"} Jan 06 15:11:26 crc kubenswrapper[4744]: I0106 15:11:26.212592 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxvg5" event={"ID":"47934586-5fc0-47bf-bae3-e7dcd356e98c","Type":"ContainerStarted","Data":"0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006"} Jan 06 15:11:26 crc kubenswrapper[4744]: I0106 15:11:26.246241 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jxvg5" podStartSLOduration=2.806049176 podStartE2EDuration="7.24614508s" podCreationTimestamp="2026-01-06 15:11:19 +0000 UTC" firstStartedPulling="2026-01-06 15:11:21.144666474 +0000 UTC m=+2077.772132792" lastFinishedPulling="2026-01-06 15:11:25.584762378 +0000 UTC m=+2082.212228696" observedRunningTime="2026-01-06 15:11:26.233537135 +0000 UTC m=+2082.861003473" watchObservedRunningTime="2026-01-06 15:11:26.24614508 +0000 UTC m=+2082.873611408" Jan 06 15:11:29 crc kubenswrapper[4744]: I0106 15:11:29.487671 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:29 crc kubenswrapper[4744]: I0106 15:11:29.488491 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:29 crc kubenswrapper[4744]: I0106 15:11:29.553365 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:30 crc kubenswrapper[4744]: I0106 15:11:30.349288 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:30 crc kubenswrapper[4744]: I0106 15:11:30.410080 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jxvg5"] Jan 06 15:11:32 crc kubenswrapper[4744]: I0106 15:11:32.294609 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jxvg5" podUID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerName="registry-server" containerID="cri-o://0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006" gracePeriod=2 Jan 06 15:11:32 crc kubenswrapper[4744]: I0106 15:11:32.919644 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.051269 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-catalog-content\") pod \"47934586-5fc0-47bf-bae3-e7dcd356e98c\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.051515 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-utilities\") pod \"47934586-5fc0-47bf-bae3-e7dcd356e98c\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.051583 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv8fz\" (UniqueName: \"kubernetes.io/projected/47934586-5fc0-47bf-bae3-e7dcd356e98c-kube-api-access-bv8fz\") pod \"47934586-5fc0-47bf-bae3-e7dcd356e98c\" (UID: \"47934586-5fc0-47bf-bae3-e7dcd356e98c\") " Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.052478 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-utilities" (OuterVolumeSpecName: "utilities") pod "47934586-5fc0-47bf-bae3-e7dcd356e98c" (UID: "47934586-5fc0-47bf-bae3-e7dcd356e98c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.059508 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47934586-5fc0-47bf-bae3-e7dcd356e98c-kube-api-access-bv8fz" (OuterVolumeSpecName: "kube-api-access-bv8fz") pod "47934586-5fc0-47bf-bae3-e7dcd356e98c" (UID: "47934586-5fc0-47bf-bae3-e7dcd356e98c"). InnerVolumeSpecName "kube-api-access-bv8fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.113419 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47934586-5fc0-47bf-bae3-e7dcd356e98c" (UID: "47934586-5fc0-47bf-bae3-e7dcd356e98c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.155111 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.155197 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv8fz\" (UniqueName: \"kubernetes.io/projected/47934586-5fc0-47bf-bae3-e7dcd356e98c-kube-api-access-bv8fz\") on node \"crc\" DevicePath \"\"" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.155228 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47934586-5fc0-47bf-bae3-e7dcd356e98c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.306458 4744 generic.go:334] "Generic (PLEG): container finished" podID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerID="0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006" exitCode=0 Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.306521 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxvg5" event={"ID":"47934586-5fc0-47bf-bae3-e7dcd356e98c","Type":"ContainerDied","Data":"0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006"} Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.306550 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jxvg5" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.306565 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jxvg5" event={"ID":"47934586-5fc0-47bf-bae3-e7dcd356e98c","Type":"ContainerDied","Data":"8a1c68dbd104f5b045b05b47b09da9e354ca985822af4e8851f5fdc952e5888f"} Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.306595 4744 scope.go:117] "RemoveContainer" containerID="0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.335015 4744 scope.go:117] "RemoveContainer" containerID="ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.358677 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jxvg5"] Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.370087 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jxvg5"] Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.378739 4744 scope.go:117] "RemoveContainer" containerID="8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.448897 4744 scope.go:117] "RemoveContainer" containerID="0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006" Jan 06 15:11:33 crc kubenswrapper[4744]: E0106 15:11:33.449488 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006\": container with ID starting with 0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006 not found: ID does not exist" containerID="0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.449542 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006"} err="failed to get container status \"0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006\": rpc error: code = NotFound desc = could not find container \"0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006\": container with ID starting with 0d1c7a839d08d7da38a51a927e4bd1d3c8a296b9bdcec714690be572c2cb0006 not found: ID does not exist" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.449575 4744 scope.go:117] "RemoveContainer" containerID="ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa" Jan 06 15:11:33 crc kubenswrapper[4744]: E0106 15:11:33.450086 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa\": container with ID starting with ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa not found: ID does not exist" containerID="ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.450128 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa"} err="failed to get container status \"ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa\": rpc error: code = NotFound desc = could not find container \"ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa\": container with ID starting with ea59dcfb580e1f0f6b7207edab1e400ffce19c738472ff6eb8ffc97ba78aa9fa not found: ID does not exist" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.450175 4744 scope.go:117] "RemoveContainer" containerID="8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37" Jan 06 15:11:33 crc kubenswrapper[4744]: E0106 15:11:33.450778 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37\": container with ID starting with 8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37 not found: ID does not exist" containerID="8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.450806 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37"} err="failed to get container status \"8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37\": rpc error: code = NotFound desc = could not find container \"8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37\": container with ID starting with 8cd01c2b23cbe74ab7c1d253c546e167fc8dd8f5d5678132509f242b5f85da37 not found: ID does not exist" Jan 06 15:11:33 crc kubenswrapper[4744]: I0106 15:11:33.733143 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47934586-5fc0-47bf-bae3-e7dcd356e98c" path="/var/lib/kubelet/pods/47934586-5fc0-47bf-bae3-e7dcd356e98c/volumes" Jan 06 15:11:54 crc kubenswrapper[4744]: I0106 15:11:54.072325 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9d6"] Jan 06 15:11:54 crc kubenswrapper[4744]: I0106 15:11:54.095755 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9d6"] Jan 06 15:11:56 crc kubenswrapper[4744]: I0106 15:11:56.137751 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eb055cf-7e1b-441b-a1db-4e6479bc9152" path="/var/lib/kubelet/pods/6eb055cf-7e1b-441b-a1db-4e6479bc9152/volumes" Jan 06 15:12:10 crc kubenswrapper[4744]: I0106 15:12:10.063954 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-pvt7q"] Jan 06 15:12:10 crc kubenswrapper[4744]: I0106 15:12:10.079139 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-pvt7q"] Jan 06 15:12:11 crc kubenswrapper[4744]: I0106 15:12:11.044700 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-4ed5-account-create-update-6k2k5"] Jan 06 15:12:11 crc kubenswrapper[4744]: I0106 15:12:11.064351 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-4ed5-account-create-update-6k2k5"] Jan 06 15:12:11 crc kubenswrapper[4744]: I0106 15:12:11.212616 4744 scope.go:117] "RemoveContainer" containerID="521ed6e2147af2111b7b6390e08ed55ffd2d09fdf8c29e971046540dd8446664" Jan 06 15:12:11 crc kubenswrapper[4744]: I0106 15:12:11.729203 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ba311c0-54e2-48cd-892b-035b7e44f2d4" path="/var/lib/kubelet/pods/7ba311c0-54e2-48cd-892b-035b7e44f2d4/volumes" Jan 06 15:12:11 crc kubenswrapper[4744]: I0106 15:12:11.731230 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3187d5d-d476-46cd-82ef-c4ce080b23ea" path="/var/lib/kubelet/pods/f3187d5d-d476-46cd-82ef-c4ce080b23ea/volumes" Jan 06 15:12:14 crc kubenswrapper[4744]: I0106 15:12:14.424710 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:12:14 crc kubenswrapper[4744]: I0106 15:12:14.425518 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:12:22 crc kubenswrapper[4744]: I0106 15:12:22.056866 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-j59t8"] Jan 06 15:12:22 crc kubenswrapper[4744]: I0106 15:12:22.077396 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-j59t8"] Jan 06 15:12:23 crc kubenswrapper[4744]: I0106 15:12:23.734104 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a102390c-a0ee-4423-91fa-b5e789fd2eac" path="/var/lib/kubelet/pods/a102390c-a0ee-4423-91fa-b5e789fd2eac/volumes" Jan 06 15:12:31 crc kubenswrapper[4744]: I0106 15:12:31.041807 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-25l7s"] Jan 06 15:12:31 crc kubenswrapper[4744]: I0106 15:12:31.053602 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-25l7s"] Jan 06 15:12:31 crc kubenswrapper[4744]: I0106 15:12:31.732543 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="780c730e-33dd-4008-958b-da44ab017991" path="/var/lib/kubelet/pods/780c730e-33dd-4008-958b-da44ab017991/volumes" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.734189 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gcxdb"] Jan 06 15:12:32 crc kubenswrapper[4744]: E0106 15:12:32.735191 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerName="extract-content" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.735219 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerName="extract-content" Jan 06 15:12:32 crc kubenswrapper[4744]: E0106 15:12:32.735287 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerName="registry-server" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.735299 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerName="registry-server" Jan 06 15:12:32 crc kubenswrapper[4744]: E0106 15:12:32.735325 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerName="extract-utilities" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.735337 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerName="extract-utilities" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.735713 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="47934586-5fc0-47bf-bae3-e7dcd356e98c" containerName="registry-server" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.738519 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.751332 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcxdb"] Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.822996 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-catalog-content\") pod \"redhat-marketplace-gcxdb\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.823252 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-utilities\") pod \"redhat-marketplace-gcxdb\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.823605 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjchg\" (UniqueName: \"kubernetes.io/projected/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-kube-api-access-fjchg\") pod \"redhat-marketplace-gcxdb\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.927486 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-utilities\") pod \"redhat-marketplace-gcxdb\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.927652 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjchg\" (UniqueName: \"kubernetes.io/projected/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-kube-api-access-fjchg\") pod \"redhat-marketplace-gcxdb\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.927725 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-catalog-content\") pod \"redhat-marketplace-gcxdb\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.928500 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-catalog-content\") pod \"redhat-marketplace-gcxdb\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.930088 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-utilities\") pod \"redhat-marketplace-gcxdb\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:32 crc kubenswrapper[4744]: I0106 15:12:32.948226 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjchg\" (UniqueName: \"kubernetes.io/projected/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-kube-api-access-fjchg\") pod \"redhat-marketplace-gcxdb\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:33 crc kubenswrapper[4744]: I0106 15:12:33.069534 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:33 crc kubenswrapper[4744]: I0106 15:12:33.605913 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcxdb"] Jan 06 15:12:34 crc kubenswrapper[4744]: I0106 15:12:34.102749 4744 generic.go:334] "Generic (PLEG): container finished" podID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerID="3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9" exitCode=0 Jan 06 15:12:34 crc kubenswrapper[4744]: I0106 15:12:34.102799 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcxdb" event={"ID":"9851ea36-bc74-4a51-8f06-7c6f6544ac9e","Type":"ContainerDied","Data":"3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9"} Jan 06 15:12:34 crc kubenswrapper[4744]: I0106 15:12:34.102847 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcxdb" event={"ID":"9851ea36-bc74-4a51-8f06-7c6f6544ac9e","Type":"ContainerStarted","Data":"c57fc953e5479bd51461af30f0430ffdd6846df09179f35c3521ea8ebd3bdffd"} Jan 06 15:12:35 crc kubenswrapper[4744]: I0106 15:12:35.116263 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcxdb" event={"ID":"9851ea36-bc74-4a51-8f06-7c6f6544ac9e","Type":"ContainerStarted","Data":"359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3"} Jan 06 15:12:36 crc kubenswrapper[4744]: I0106 15:12:36.135209 4744 generic.go:334] "Generic (PLEG): container finished" podID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerID="359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3" exitCode=0 Jan 06 15:12:36 crc kubenswrapper[4744]: I0106 15:12:36.135665 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcxdb" event={"ID":"9851ea36-bc74-4a51-8f06-7c6f6544ac9e","Type":"ContainerDied","Data":"359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3"} Jan 06 15:12:37 crc kubenswrapper[4744]: I0106 15:12:37.154217 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcxdb" event={"ID":"9851ea36-bc74-4a51-8f06-7c6f6544ac9e","Type":"ContainerStarted","Data":"a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db"} Jan 06 15:12:37 crc kubenswrapper[4744]: I0106 15:12:37.184892 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gcxdb" podStartSLOduration=2.539715321 podStartE2EDuration="5.184871448s" podCreationTimestamp="2026-01-06 15:12:32 +0000 UTC" firstStartedPulling="2026-01-06 15:12:34.106686219 +0000 UTC m=+2150.734152537" lastFinishedPulling="2026-01-06 15:12:36.751842316 +0000 UTC m=+2153.379308664" observedRunningTime="2026-01-06 15:12:37.179772102 +0000 UTC m=+2153.807238450" watchObservedRunningTime="2026-01-06 15:12:37.184871448 +0000 UTC m=+2153.812337766" Jan 06 15:12:43 crc kubenswrapper[4744]: I0106 15:12:43.070303 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:43 crc kubenswrapper[4744]: I0106 15:12:43.070862 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:43 crc kubenswrapper[4744]: I0106 15:12:43.148955 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:43 crc kubenswrapper[4744]: I0106 15:12:43.294767 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:43 crc kubenswrapper[4744]: I0106 15:12:43.398912 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcxdb"] Jan 06 15:12:44 crc kubenswrapper[4744]: I0106 15:12:44.424422 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:12:44 crc kubenswrapper[4744]: I0106 15:12:44.424491 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.246150 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gcxdb" podUID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerName="registry-server" containerID="cri-o://a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db" gracePeriod=2 Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.738628 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.785550 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-catalog-content\") pod \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.785793 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjchg\" (UniqueName: \"kubernetes.io/projected/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-kube-api-access-fjchg\") pod \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.785851 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-utilities\") pod \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\" (UID: \"9851ea36-bc74-4a51-8f06-7c6f6544ac9e\") " Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.787598 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-utilities" (OuterVolumeSpecName: "utilities") pod "9851ea36-bc74-4a51-8f06-7c6f6544ac9e" (UID: "9851ea36-bc74-4a51-8f06-7c6f6544ac9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.793666 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-kube-api-access-fjchg" (OuterVolumeSpecName: "kube-api-access-fjchg") pod "9851ea36-bc74-4a51-8f06-7c6f6544ac9e" (UID: "9851ea36-bc74-4a51-8f06-7c6f6544ac9e"). InnerVolumeSpecName "kube-api-access-fjchg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.818290 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9851ea36-bc74-4a51-8f06-7c6f6544ac9e" (UID: "9851ea36-bc74-4a51-8f06-7c6f6544ac9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.889285 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.889315 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjchg\" (UniqueName: \"kubernetes.io/projected/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-kube-api-access-fjchg\") on node \"crc\" DevicePath \"\"" Jan 06 15:12:45 crc kubenswrapper[4744]: I0106 15:12:45.889326 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9851ea36-bc74-4a51-8f06-7c6f6544ac9e-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.260932 4744 generic.go:334] "Generic (PLEG): container finished" podID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerID="a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db" exitCode=0 Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.261007 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcxdb" event={"ID":"9851ea36-bc74-4a51-8f06-7c6f6544ac9e","Type":"ContainerDied","Data":"a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db"} Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.261341 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gcxdb" event={"ID":"9851ea36-bc74-4a51-8f06-7c6f6544ac9e","Type":"ContainerDied","Data":"c57fc953e5479bd51461af30f0430ffdd6846df09179f35c3521ea8ebd3bdffd"} Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.261372 4744 scope.go:117] "RemoveContainer" containerID="a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.261059 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gcxdb" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.301951 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcxdb"] Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.302150 4744 scope.go:117] "RemoveContainer" containerID="359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.313008 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gcxdb"] Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.325418 4744 scope.go:117] "RemoveContainer" containerID="3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.408224 4744 scope.go:117] "RemoveContainer" containerID="a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db" Jan 06 15:12:46 crc kubenswrapper[4744]: E0106 15:12:46.408749 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db\": container with ID starting with a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db not found: ID does not exist" containerID="a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.408810 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db"} err="failed to get container status \"a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db\": rpc error: code = NotFound desc = could not find container \"a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db\": container with ID starting with a28b45cf42a2744ae5865f372807bd9934e8597e94f946340996092be971b8db not found: ID does not exist" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.408849 4744 scope.go:117] "RemoveContainer" containerID="359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3" Jan 06 15:12:46 crc kubenswrapper[4744]: E0106 15:12:46.409266 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3\": container with ID starting with 359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3 not found: ID does not exist" containerID="359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.409296 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3"} err="failed to get container status \"359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3\": rpc error: code = NotFound desc = could not find container \"359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3\": container with ID starting with 359338f72243844ccda6c13ac4f1d612c6797dc409c03152c9d002e0bb56d3e3 not found: ID does not exist" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.409319 4744 scope.go:117] "RemoveContainer" containerID="3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9" Jan 06 15:12:46 crc kubenswrapper[4744]: E0106 15:12:46.409563 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9\": container with ID starting with 3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9 not found: ID does not exist" containerID="3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9" Jan 06 15:12:46 crc kubenswrapper[4744]: I0106 15:12:46.409600 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9"} err="failed to get container status \"3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9\": rpc error: code = NotFound desc = could not find container \"3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9\": container with ID starting with 3451bcecbb0b20e966f3e2897637e36638bcfb4972e1e8116ebbeeac7afd3fd9 not found: ID does not exist" Jan 06 15:12:47 crc kubenswrapper[4744]: I0106 15:12:47.735852 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" path="/var/lib/kubelet/pods/9851ea36-bc74-4a51-8f06-7c6f6544ac9e/volumes" Jan 06 15:13:11 crc kubenswrapper[4744]: I0106 15:13:11.365420 4744 scope.go:117] "RemoveContainer" containerID="83f8c07d400a3b6b77f900ef57d64cc4443225d3c862be1834309544e20b77dc" Jan 06 15:13:11 crc kubenswrapper[4744]: I0106 15:13:11.414691 4744 scope.go:117] "RemoveContainer" containerID="d9413cf78bfc30666af986f6257ca8809ed298701060bc9bc879d08c6a3f6468" Jan 06 15:13:11 crc kubenswrapper[4744]: I0106 15:13:11.457363 4744 scope.go:117] "RemoveContainer" containerID="904d6f3f7a389ffaf5a5cd0e1fed71c2ff096f7c520bc3dba8c4b628823a5fb1" Jan 06 15:13:11 crc kubenswrapper[4744]: I0106 15:13:11.512191 4744 scope.go:117] "RemoveContainer" containerID="4cd4b0933ac9f7630420095487c9fa00a431eece7059c4011db386c7446f3d54" Jan 06 15:13:14 crc kubenswrapper[4744]: I0106 15:13:14.424282 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:13:14 crc kubenswrapper[4744]: I0106 15:13:14.424876 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:13:14 crc kubenswrapper[4744]: I0106 15:13:14.424938 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:13:14 crc kubenswrapper[4744]: I0106 15:13:14.426041 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:13:14 crc kubenswrapper[4744]: I0106 15:13:14.426146 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" gracePeriod=600 Jan 06 15:13:14 crc kubenswrapper[4744]: E0106 15:13:14.557827 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:13:14 crc kubenswrapper[4744]: I0106 15:13:14.621413 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" exitCode=0 Jan 06 15:13:14 crc kubenswrapper[4744]: I0106 15:13:14.621477 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce"} Jan 06 15:13:14 crc kubenswrapper[4744]: I0106 15:13:14.621522 4744 scope.go:117] "RemoveContainer" containerID="8d70e5126160c45c7349d68db9fd09dd06c15ea7a0105b790d710f8a5423e632" Jan 06 15:13:14 crc kubenswrapper[4744]: I0106 15:13:14.623559 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:13:14 crc kubenswrapper[4744]: E0106 15:13:14.624062 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:13:16 crc kubenswrapper[4744]: I0106 15:13:16.063627 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-8gfkz"] Jan 06 15:13:16 crc kubenswrapper[4744]: I0106 15:13:16.078952 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-8gfkz"] Jan 06 15:13:17 crc kubenswrapper[4744]: I0106 15:13:17.745890 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75eb54e3-63f3-4766-b95b-dbb1b307950c" path="/var/lib/kubelet/pods/75eb54e3-63f3-4766-b95b-dbb1b307950c/volumes" Jan 06 15:13:23 crc kubenswrapper[4744]: I0106 15:13:23.750428 4744 generic.go:334] "Generic (PLEG): container finished" podID="98606045-db2f-49f7-a0a3-c88b391364c6" containerID="cbfe473dca0173764cd2dd40d74844ed84c7312db4282d71da0c653dc8f19894" exitCode=0 Jan 06 15:13:23 crc kubenswrapper[4744]: I0106 15:13:23.750534 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" event={"ID":"98606045-db2f-49f7-a0a3-c88b391364c6","Type":"ContainerDied","Data":"cbfe473dca0173764cd2dd40d74844ed84c7312db4282d71da0c653dc8f19894"} Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.417979 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.498559 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvpzs\" (UniqueName: \"kubernetes.io/projected/98606045-db2f-49f7-a0a3-c88b391364c6-kube-api-access-mvpzs\") pod \"98606045-db2f-49f7-a0a3-c88b391364c6\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.498724 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-inventory\") pod \"98606045-db2f-49f7-a0a3-c88b391364c6\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.498779 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-ssh-key-openstack-edpm-ipam\") pod \"98606045-db2f-49f7-a0a3-c88b391364c6\" (UID: \"98606045-db2f-49f7-a0a3-c88b391364c6\") " Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.505267 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98606045-db2f-49f7-a0a3-c88b391364c6-kube-api-access-mvpzs" (OuterVolumeSpecName: "kube-api-access-mvpzs") pod "98606045-db2f-49f7-a0a3-c88b391364c6" (UID: "98606045-db2f-49f7-a0a3-c88b391364c6"). InnerVolumeSpecName "kube-api-access-mvpzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.536960 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "98606045-db2f-49f7-a0a3-c88b391364c6" (UID: "98606045-db2f-49f7-a0a3-c88b391364c6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.548980 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-inventory" (OuterVolumeSpecName: "inventory") pod "98606045-db2f-49f7-a0a3-c88b391364c6" (UID: "98606045-db2f-49f7-a0a3-c88b391364c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.601960 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvpzs\" (UniqueName: \"kubernetes.io/projected/98606045-db2f-49f7-a0a3-c88b391364c6-kube-api-access-mvpzs\") on node \"crc\" DevicePath \"\"" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.602000 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.602013 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/98606045-db2f-49f7-a0a3-c88b391364c6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.797837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" event={"ID":"98606045-db2f-49f7-a0a3-c88b391364c6","Type":"ContainerDied","Data":"45c421d0e1e5cdd9358fa497a6db4f71dcd7981850eefdb994b618101bfaedd4"} Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.797938 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45c421d0e1e5cdd9358fa497a6db4f71dcd7981850eefdb994b618101bfaedd4" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.797967 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.897840 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9"] Jan 06 15:13:25 crc kubenswrapper[4744]: E0106 15:13:25.898674 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerName="extract-content" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.898705 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerName="extract-content" Jan 06 15:13:25 crc kubenswrapper[4744]: E0106 15:13:25.898785 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerName="registry-server" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.898799 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerName="registry-server" Jan 06 15:13:25 crc kubenswrapper[4744]: E0106 15:13:25.898825 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98606045-db2f-49f7-a0a3-c88b391364c6" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.898840 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="98606045-db2f-49f7-a0a3-c88b391364c6" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 06 15:13:25 crc kubenswrapper[4744]: E0106 15:13:25.898864 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerName="extract-utilities" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.898879 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerName="extract-utilities" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.899342 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="9851ea36-bc74-4a51-8f06-7c6f6544ac9e" containerName="registry-server" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.899382 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="98606045-db2f-49f7-a0a3-c88b391364c6" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.900751 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.904693 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.905440 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.905558 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.905662 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:13:25 crc kubenswrapper[4744]: I0106 15:13:25.926815 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9"] Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.037940 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.038115 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fxzb\" (UniqueName: \"kubernetes.io/projected/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-kube-api-access-4fxzb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.038230 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.141197 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.141331 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fxzb\" (UniqueName: \"kubernetes.io/projected/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-kube-api-access-4fxzb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.141400 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.145748 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.150709 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.176138 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fxzb\" (UniqueName: \"kubernetes.io/projected/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-kube-api-access-4fxzb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.230044 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:13:26 crc kubenswrapper[4744]: I0106 15:13:26.835146 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9"] Jan 06 15:13:27 crc kubenswrapper[4744]: I0106 15:13:27.834541 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" event={"ID":"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55","Type":"ContainerStarted","Data":"6e9e54f55bba28b2726ae4d46cff0a816f7f26639e716c7a7f8e84541504e59e"} Jan 06 15:13:27 crc kubenswrapper[4744]: I0106 15:13:27.835333 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" event={"ID":"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55","Type":"ContainerStarted","Data":"869445bd0ef0ad3e5cfdb2376a74c694f1bc71362a9ad468f0b2fe44925ccf56"} Jan 06 15:13:27 crc kubenswrapper[4744]: I0106 15:13:27.857188 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" podStartSLOduration=2.22934517 podStartE2EDuration="2.85715115s" podCreationTimestamp="2026-01-06 15:13:25 +0000 UTC" firstStartedPulling="2026-01-06 15:13:26.834076792 +0000 UTC m=+2203.461543120" lastFinishedPulling="2026-01-06 15:13:27.461882742 +0000 UTC m=+2204.089349100" observedRunningTime="2026-01-06 15:13:27.85225116 +0000 UTC m=+2204.479717528" watchObservedRunningTime="2026-01-06 15:13:27.85715115 +0000 UTC m=+2204.484617468" Jan 06 15:13:28 crc kubenswrapper[4744]: I0106 15:13:28.711758 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:13:28 crc kubenswrapper[4744]: E0106 15:13:28.712500 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:13:43 crc kubenswrapper[4744]: I0106 15:13:43.738554 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:13:43 crc kubenswrapper[4744]: E0106 15:13:43.740835 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:13:55 crc kubenswrapper[4744]: I0106 15:13:55.711819 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:13:55 crc kubenswrapper[4744]: E0106 15:13:55.712707 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:14:10 crc kubenswrapper[4744]: I0106 15:14:10.711712 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:14:10 crc kubenswrapper[4744]: E0106 15:14:10.712752 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:14:11 crc kubenswrapper[4744]: I0106 15:14:11.695933 4744 scope.go:117] "RemoveContainer" containerID="d268bae99d5a17ddfa477ccd0a637090326bbeff11dba3c3456b8a3579abdc86" Jan 06 15:14:22 crc kubenswrapper[4744]: I0106 15:14:22.713029 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:14:22 crc kubenswrapper[4744]: E0106 15:14:22.714863 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:14:37 crc kubenswrapper[4744]: I0106 15:14:37.712151 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:14:37 crc kubenswrapper[4744]: E0106 15:14:37.713399 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:14:49 crc kubenswrapper[4744]: I0106 15:14:49.711291 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:14:49 crc kubenswrapper[4744]: E0106 15:14:49.712264 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:14:52 crc kubenswrapper[4744]: I0106 15:14:52.046084 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-ptgpq"] Jan 06 15:14:52 crc kubenswrapper[4744]: I0106 15:14:52.056529 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-ptgpq"] Jan 06 15:14:52 crc kubenswrapper[4744]: I0106 15:14:52.073551 4744 generic.go:334] "Generic (PLEG): container finished" podID="ae6684c3-9c28-4aca-a8c3-7a4545d6eb55" containerID="6e9e54f55bba28b2726ae4d46cff0a816f7f26639e716c7a7f8e84541504e59e" exitCode=0 Jan 06 15:14:52 crc kubenswrapper[4744]: I0106 15:14:52.073597 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" event={"ID":"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55","Type":"ContainerDied","Data":"6e9e54f55bba28b2726ae4d46cff0a816f7f26639e716c7a7f8e84541504e59e"} Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.659123 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.731650 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dcc5556-fb95-4735-b08c-a11e0e2d7042" path="/var/lib/kubelet/pods/0dcc5556-fb95-4735-b08c-a11e0e2d7042/volumes" Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.797598 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fxzb\" (UniqueName: \"kubernetes.io/projected/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-kube-api-access-4fxzb\") pod \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.797944 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-ssh-key-openstack-edpm-ipam\") pod \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.798083 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-inventory\") pod \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\" (UID: \"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55\") " Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.806456 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-kube-api-access-4fxzb" (OuterVolumeSpecName: "kube-api-access-4fxzb") pod "ae6684c3-9c28-4aca-a8c3-7a4545d6eb55" (UID: "ae6684c3-9c28-4aca-a8c3-7a4545d6eb55"). InnerVolumeSpecName "kube-api-access-4fxzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.832404 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-inventory" (OuterVolumeSpecName: "inventory") pod "ae6684c3-9c28-4aca-a8c3-7a4545d6eb55" (UID: "ae6684c3-9c28-4aca-a8c3-7a4545d6eb55"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.838705 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ae6684c3-9c28-4aca-a8c3-7a4545d6eb55" (UID: "ae6684c3-9c28-4aca-a8c3-7a4545d6eb55"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.916870 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fxzb\" (UniqueName: \"kubernetes.io/projected/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-kube-api-access-4fxzb\") on node \"crc\" DevicePath \"\"" Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.916905 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:14:53 crc kubenswrapper[4744]: I0106 15:14:53.916923 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae6684c3-9c28-4aca-a8c3-7a4545d6eb55-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.100326 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" event={"ID":"ae6684c3-9c28-4aca-a8c3-7a4545d6eb55","Type":"ContainerDied","Data":"869445bd0ef0ad3e5cfdb2376a74c694f1bc71362a9ad468f0b2fe44925ccf56"} Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.100368 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="869445bd0ef0ad3e5cfdb2376a74c694f1bc71362a9ad468f0b2fe44925ccf56" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.100410 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.194545 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9"] Jan 06 15:14:54 crc kubenswrapper[4744]: E0106 15:14:54.195292 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae6684c3-9c28-4aca-a8c3-7a4545d6eb55" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.195316 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae6684c3-9c28-4aca-a8c3-7a4545d6eb55" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.195644 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae6684c3-9c28-4aca-a8c3-7a4545d6eb55" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.196686 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.201302 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.201495 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.201668 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.201814 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.205061 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9"] Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.325863 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.326292 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.326891 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lv67\" (UniqueName: \"kubernetes.io/projected/f0380bc6-d687-41af-a5d5-6569a6c299d9-kube-api-access-5lv67\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.429126 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lv67\" (UniqueName: \"kubernetes.io/projected/f0380bc6-d687-41af-a5d5-6569a6c299d9-kube-api-access-5lv67\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.429218 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.429333 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.432889 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.440506 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.450404 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lv67\" (UniqueName: \"kubernetes.io/projected/f0380bc6-d687-41af-a5d5-6569a6c299d9-kube-api-access-5lv67\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:54 crc kubenswrapper[4744]: I0106 15:14:54.522125 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:14:55 crc kubenswrapper[4744]: I0106 15:14:55.124945 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9"] Jan 06 15:14:55 crc kubenswrapper[4744]: W0106 15:14:55.137096 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0380bc6_d687_41af_a5d5_6569a6c299d9.slice/crio-e372b764becac99ad640ba29b8e5a680be53c003b224101d9f46421b53cd5d5c WatchSource:0}: Error finding container e372b764becac99ad640ba29b8e5a680be53c003b224101d9f46421b53cd5d5c: Status 404 returned error can't find the container with id e372b764becac99ad640ba29b8e5a680be53c003b224101d9f46421b53cd5d5c Jan 06 15:14:55 crc kubenswrapper[4744]: I0106 15:14:55.141279 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 15:14:56 crc kubenswrapper[4744]: I0106 15:14:56.147258 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" event={"ID":"f0380bc6-d687-41af-a5d5-6569a6c299d9","Type":"ContainerStarted","Data":"e372b764becac99ad640ba29b8e5a680be53c003b224101d9f46421b53cd5d5c"} Jan 06 15:14:57 crc kubenswrapper[4744]: I0106 15:14:57.161283 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" event={"ID":"f0380bc6-d687-41af-a5d5-6569a6c299d9","Type":"ContainerStarted","Data":"13dedfb0e0ce2367c4c9c19ad84bdf631b39b616efcd49a9fb885b6ad20c63b9"} Jan 06 15:14:57 crc kubenswrapper[4744]: I0106 15:14:57.211786 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" podStartSLOduration=2.277805441 podStartE2EDuration="3.211751398s" podCreationTimestamp="2026-01-06 15:14:54 +0000 UTC" firstStartedPulling="2026-01-06 15:14:55.141022521 +0000 UTC m=+2291.768488849" lastFinishedPulling="2026-01-06 15:14:56.074968478 +0000 UTC m=+2292.702434806" observedRunningTime="2026-01-06 15:14:57.194588512 +0000 UTC m=+2293.822054830" watchObservedRunningTime="2026-01-06 15:14:57.211751398 +0000 UTC m=+2293.839217756" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.152112 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs"] Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.156389 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.158517 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.159024 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.166544 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs"] Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.288950 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbwbk\" (UniqueName: \"kubernetes.io/projected/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-kube-api-access-rbwbk\") pod \"collect-profiles-29461875-z62cs\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.289267 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-config-volume\") pod \"collect-profiles-29461875-z62cs\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.289454 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-secret-volume\") pod \"collect-profiles-29461875-z62cs\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.391243 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbwbk\" (UniqueName: \"kubernetes.io/projected/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-kube-api-access-rbwbk\") pod \"collect-profiles-29461875-z62cs\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.391291 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-config-volume\") pod \"collect-profiles-29461875-z62cs\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.391479 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-secret-volume\") pod \"collect-profiles-29461875-z62cs\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.392310 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-config-volume\") pod \"collect-profiles-29461875-z62cs\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.397084 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-secret-volume\") pod \"collect-profiles-29461875-z62cs\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.411325 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbwbk\" (UniqueName: \"kubernetes.io/projected/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-kube-api-access-rbwbk\") pod \"collect-profiles-29461875-z62cs\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.502690 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:00 crc kubenswrapper[4744]: I0106 15:15:00.991900 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs"] Jan 06 15:15:01 crc kubenswrapper[4744]: I0106 15:15:01.228330 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" event={"ID":"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c","Type":"ContainerStarted","Data":"c5366a46dc52c1062676c4b450fa191f3b76a427236e1fff7cee7de78c703970"} Jan 06 15:15:01 crc kubenswrapper[4744]: I0106 15:15:01.228653 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" event={"ID":"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c","Type":"ContainerStarted","Data":"402aee6b9cd765d44480ae4847612c66f0d7a619154c9093a1fb62d9293cbcc5"} Jan 06 15:15:01 crc kubenswrapper[4744]: I0106 15:15:01.246526 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" podStartSLOduration=1.246508526 podStartE2EDuration="1.246508526s" podCreationTimestamp="2026-01-06 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:15:01.244611406 +0000 UTC m=+2297.872077724" watchObservedRunningTime="2026-01-06 15:15:01.246508526 +0000 UTC m=+2297.873974854" Jan 06 15:15:01 crc kubenswrapper[4744]: I0106 15:15:01.711813 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:15:01 crc kubenswrapper[4744]: E0106 15:15:01.712123 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:15:02 crc kubenswrapper[4744]: I0106 15:15:02.244078 4744 generic.go:334] "Generic (PLEG): container finished" podID="f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c" containerID="c5366a46dc52c1062676c4b450fa191f3b76a427236e1fff7cee7de78c703970" exitCode=0 Jan 06 15:15:02 crc kubenswrapper[4744]: I0106 15:15:02.244242 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" event={"ID":"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c","Type":"ContainerDied","Data":"c5366a46dc52c1062676c4b450fa191f3b76a427236e1fff7cee7de78c703970"} Jan 06 15:15:02 crc kubenswrapper[4744]: I0106 15:15:02.248672 4744 generic.go:334] "Generic (PLEG): container finished" podID="f0380bc6-d687-41af-a5d5-6569a6c299d9" containerID="13dedfb0e0ce2367c4c9c19ad84bdf631b39b616efcd49a9fb885b6ad20c63b9" exitCode=0 Jan 06 15:15:02 crc kubenswrapper[4744]: I0106 15:15:02.248717 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" event={"ID":"f0380bc6-d687-41af-a5d5-6569a6c299d9","Type":"ContainerDied","Data":"13dedfb0e0ce2367c4c9c19ad84bdf631b39b616efcd49a9fb885b6ad20c63b9"} Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.711695 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.794251 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-secret-volume\") pod \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.794502 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbwbk\" (UniqueName: \"kubernetes.io/projected/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-kube-api-access-rbwbk\") pod \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.794535 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-config-volume\") pod \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\" (UID: \"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c\") " Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.795885 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-config-volume" (OuterVolumeSpecName: "config-volume") pod "f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c" (UID: "f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.807536 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-kube-api-access-rbwbk" (OuterVolumeSpecName: "kube-api-access-rbwbk") pod "f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c" (UID: "f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c"). InnerVolumeSpecName "kube-api-access-rbwbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.808120 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c" (UID: "f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.883259 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.899477 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbwbk\" (UniqueName: \"kubernetes.io/projected/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-kube-api-access-rbwbk\") on node \"crc\" DevicePath \"\"" Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.899508 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 15:15:03 crc kubenswrapper[4744]: I0106 15:15:03.899517 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.001009 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lv67\" (UniqueName: \"kubernetes.io/projected/f0380bc6-d687-41af-a5d5-6569a6c299d9-kube-api-access-5lv67\") pod \"f0380bc6-d687-41af-a5d5-6569a6c299d9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.001200 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-ssh-key-openstack-edpm-ipam\") pod \"f0380bc6-d687-41af-a5d5-6569a6c299d9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.001313 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-inventory\") pod \"f0380bc6-d687-41af-a5d5-6569a6c299d9\" (UID: \"f0380bc6-d687-41af-a5d5-6569a6c299d9\") " Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.004697 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0380bc6-d687-41af-a5d5-6569a6c299d9-kube-api-access-5lv67" (OuterVolumeSpecName: "kube-api-access-5lv67") pod "f0380bc6-d687-41af-a5d5-6569a6c299d9" (UID: "f0380bc6-d687-41af-a5d5-6569a6c299d9"). InnerVolumeSpecName "kube-api-access-5lv67". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.029905 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-inventory" (OuterVolumeSpecName: "inventory") pod "f0380bc6-d687-41af-a5d5-6569a6c299d9" (UID: "f0380bc6-d687-41af-a5d5-6569a6c299d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.054693 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f0380bc6-d687-41af-a5d5-6569a6c299d9" (UID: "f0380bc6-d687-41af-a5d5-6569a6c299d9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.104433 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.104471 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0380bc6-d687-41af-a5d5-6569a6c299d9-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.104480 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lv67\" (UniqueName: \"kubernetes.io/projected/f0380bc6-d687-41af-a5d5-6569a6c299d9-kube-api-access-5lv67\") on node \"crc\" DevicePath \"\"" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.284017 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" event={"ID":"f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c","Type":"ContainerDied","Data":"402aee6b9cd765d44480ae4847612c66f0d7a619154c9093a1fb62d9293cbcc5"} Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.284051 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="402aee6b9cd765d44480ae4847612c66f0d7a619154c9093a1fb62d9293cbcc5" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.284116 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.303202 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" event={"ID":"f0380bc6-d687-41af-a5d5-6569a6c299d9","Type":"ContainerDied","Data":"e372b764becac99ad640ba29b8e5a680be53c003b224101d9f46421b53cd5d5c"} Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.303256 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e372b764becac99ad640ba29b8e5a680be53c003b224101d9f46421b53cd5d5c" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.303325 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.372036 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm"] Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.388694 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461830-7v4gm"] Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.489533 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb"] Jan 06 15:15:04 crc kubenswrapper[4744]: E0106 15:15:04.490100 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0380bc6-d687-41af-a5d5-6569a6c299d9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.490129 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0380bc6-d687-41af-a5d5-6569a6c299d9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 06 15:15:04 crc kubenswrapper[4744]: E0106 15:15:04.490140 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c" containerName="collect-profiles" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.490147 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c" containerName="collect-profiles" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.490453 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c" containerName="collect-profiles" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.490488 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0380bc6-d687-41af-a5d5-6569a6c299d9" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.491543 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.494507 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.495808 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.500789 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.500817 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.518391 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb"] Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.616906 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jb2sb\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.617515 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fh25\" (UniqueName: \"kubernetes.io/projected/f50fabe8-fbbb-406f-9786-ff32ff95feaa-kube-api-access-7fh25\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jb2sb\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.617737 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jb2sb\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.719225 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jb2sb\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.719336 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fh25\" (UniqueName: \"kubernetes.io/projected/f50fabe8-fbbb-406f-9786-ff32ff95feaa-kube-api-access-7fh25\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jb2sb\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.719559 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jb2sb\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.725469 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jb2sb\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.725484 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jb2sb\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.746513 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fh25\" (UniqueName: \"kubernetes.io/projected/f50fabe8-fbbb-406f-9786-ff32ff95feaa-kube-api-access-7fh25\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-jb2sb\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:04 crc kubenswrapper[4744]: I0106 15:15:04.816878 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:05 crc kubenswrapper[4744]: W0106 15:15:05.440900 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf50fabe8_fbbb_406f_9786_ff32ff95feaa.slice/crio-ce9efa2adc6a96b29ae556d479029fb7ea61e29c028fa265589ba84036ba1d62 WatchSource:0}: Error finding container ce9efa2adc6a96b29ae556d479029fb7ea61e29c028fa265589ba84036ba1d62: Status 404 returned error can't find the container with id ce9efa2adc6a96b29ae556d479029fb7ea61e29c028fa265589ba84036ba1d62 Jan 06 15:15:05 crc kubenswrapper[4744]: I0106 15:15:05.446200 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb"] Jan 06 15:15:05 crc kubenswrapper[4744]: I0106 15:15:05.731855 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e" path="/var/lib/kubelet/pods/3b23a3cd-0a2e-4815-aba8-bae5cb74fa7e/volumes" Jan 06 15:15:06 crc kubenswrapper[4744]: I0106 15:15:06.325096 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" event={"ID":"f50fabe8-fbbb-406f-9786-ff32ff95feaa","Type":"ContainerStarted","Data":"ce9efa2adc6a96b29ae556d479029fb7ea61e29c028fa265589ba84036ba1d62"} Jan 06 15:15:07 crc kubenswrapper[4744]: I0106 15:15:07.336133 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" event={"ID":"f50fabe8-fbbb-406f-9786-ff32ff95feaa","Type":"ContainerStarted","Data":"16b7e13f8027e6f1abd87866a33a8aadcf1f36c075ea0b4426c25660af158fe8"} Jan 06 15:15:07 crc kubenswrapper[4744]: I0106 15:15:07.355318 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" podStartSLOduration=2.8179073040000002 podStartE2EDuration="3.35528808s" podCreationTimestamp="2026-01-06 15:15:04 +0000 UTC" firstStartedPulling="2026-01-06 15:15:05.445632494 +0000 UTC m=+2302.073098822" lastFinishedPulling="2026-01-06 15:15:05.98301327 +0000 UTC m=+2302.610479598" observedRunningTime="2026-01-06 15:15:07.352182037 +0000 UTC m=+2303.979648355" watchObservedRunningTime="2026-01-06 15:15:07.35528808 +0000 UTC m=+2303.982754438" Jan 06 15:15:11 crc kubenswrapper[4744]: I0106 15:15:11.783767 4744 scope.go:117] "RemoveContainer" containerID="5e1d7f929690066668c78fb2b00db034662be46a2b3f34992dcc7ff404ad3ec6" Jan 06 15:15:11 crc kubenswrapper[4744]: I0106 15:15:11.825401 4744 scope.go:117] "RemoveContainer" containerID="16c5ca1f08ffad2f6c4a5f3a7a25e8fcb6bd34d9398d5bed71783bfdd8854ea8" Jan 06 15:15:15 crc kubenswrapper[4744]: I0106 15:15:15.711976 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:15:15 crc kubenswrapper[4744]: E0106 15:15:15.713144 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:15:27 crc kubenswrapper[4744]: I0106 15:15:27.712231 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:15:27 crc kubenswrapper[4744]: E0106 15:15:27.713192 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:15:39 crc kubenswrapper[4744]: I0106 15:15:39.711979 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:15:39 crc kubenswrapper[4744]: E0106 15:15:39.712752 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:15:48 crc kubenswrapper[4744]: E0106 15:15:48.278629 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf50fabe8_fbbb_406f_9786_ff32ff95feaa.slice/crio-16b7e13f8027e6f1abd87866a33a8aadcf1f36c075ea0b4426c25660af158fe8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf50fabe8_fbbb_406f_9786_ff32ff95feaa.slice/crio-conmon-16b7e13f8027e6f1abd87866a33a8aadcf1f36c075ea0b4426c25660af158fe8.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:15:48 crc kubenswrapper[4744]: E0106 15:15:48.278630 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf50fabe8_fbbb_406f_9786_ff32ff95feaa.slice/crio-16b7e13f8027e6f1abd87866a33a8aadcf1f36c075ea0b4426c25660af158fe8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf50fabe8_fbbb_406f_9786_ff32ff95feaa.slice/crio-conmon-16b7e13f8027e6f1abd87866a33a8aadcf1f36c075ea0b4426c25660af158fe8.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:15:48 crc kubenswrapper[4744]: I0106 15:15:48.839985 4744 generic.go:334] "Generic (PLEG): container finished" podID="f50fabe8-fbbb-406f-9786-ff32ff95feaa" containerID="16b7e13f8027e6f1abd87866a33a8aadcf1f36c075ea0b4426c25660af158fe8" exitCode=0 Jan 06 15:15:48 crc kubenswrapper[4744]: I0106 15:15:48.840060 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" event={"ID":"f50fabe8-fbbb-406f-9786-ff32ff95feaa","Type":"ContainerDied","Data":"16b7e13f8027e6f1abd87866a33a8aadcf1f36c075ea0b4426c25660af158fe8"} Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.348604 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.469975 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fh25\" (UniqueName: \"kubernetes.io/projected/f50fabe8-fbbb-406f-9786-ff32ff95feaa-kube-api-access-7fh25\") pod \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.470371 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-inventory\") pod \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.470541 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-ssh-key-openstack-edpm-ipam\") pod \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\" (UID: \"f50fabe8-fbbb-406f-9786-ff32ff95feaa\") " Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.480770 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f50fabe8-fbbb-406f-9786-ff32ff95feaa-kube-api-access-7fh25" (OuterVolumeSpecName: "kube-api-access-7fh25") pod "f50fabe8-fbbb-406f-9786-ff32ff95feaa" (UID: "f50fabe8-fbbb-406f-9786-ff32ff95feaa"). InnerVolumeSpecName "kube-api-access-7fh25". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.522287 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f50fabe8-fbbb-406f-9786-ff32ff95feaa" (UID: "f50fabe8-fbbb-406f-9786-ff32ff95feaa"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.524038 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-inventory" (OuterVolumeSpecName: "inventory") pod "f50fabe8-fbbb-406f-9786-ff32ff95feaa" (UID: "f50fabe8-fbbb-406f-9786-ff32ff95feaa"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.573577 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.573626 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fh25\" (UniqueName: \"kubernetes.io/projected/f50fabe8-fbbb-406f-9786-ff32ff95feaa-kube-api-access-7fh25\") on node \"crc\" DevicePath \"\"" Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.573637 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50fabe8-fbbb-406f-9786-ff32ff95feaa-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.874697 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" event={"ID":"f50fabe8-fbbb-406f-9786-ff32ff95feaa","Type":"ContainerDied","Data":"ce9efa2adc6a96b29ae556d479029fb7ea61e29c028fa265589ba84036ba1d62"} Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.874756 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce9efa2adc6a96b29ae556d479029fb7ea61e29c028fa265589ba84036ba1d62" Jan 06 15:15:50 crc kubenswrapper[4744]: I0106 15:15:50.874834 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-jb2sb" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.011610 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4"] Jan 06 15:15:51 crc kubenswrapper[4744]: E0106 15:15:51.012274 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f50fabe8-fbbb-406f-9786-ff32ff95feaa" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.012297 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f50fabe8-fbbb-406f-9786-ff32ff95feaa" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.012606 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f50fabe8-fbbb-406f-9786-ff32ff95feaa" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.013744 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.016524 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.016906 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.016910 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.017442 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.043609 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4"] Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.197834 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.197982 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpp7d\" (UniqueName: \"kubernetes.io/projected/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-kube-api-access-dpp7d\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.200183 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.303040 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.303625 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpp7d\" (UniqueName: \"kubernetes.io/projected/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-kube-api-access-dpp7d\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.303965 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.311814 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.324107 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.331783 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpp7d\" (UniqueName: \"kubernetes.io/projected/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-kube-api-access-dpp7d\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.347739 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.711406 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:15:51 crc kubenswrapper[4744]: E0106 15:15:51.711981 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:15:51 crc kubenswrapper[4744]: I0106 15:15:51.937997 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4"] Jan 06 15:15:52 crc kubenswrapper[4744]: I0106 15:15:52.912759 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" event={"ID":"2b9ab759-a383-4ec8-ad5a-578f35de2a5e","Type":"ContainerStarted","Data":"3b7ce69223c3dad14ccf3dec41edb75d5ebbfa653c57fa7d5d7ef0311fb0f75c"} Jan 06 15:15:52 crc kubenswrapper[4744]: I0106 15:15:52.913391 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" event={"ID":"2b9ab759-a383-4ec8-ad5a-578f35de2a5e","Type":"ContainerStarted","Data":"ca3de53e95b29b7a12cde9c5591f2fb2dc667c49d9fda14e058cb802db67da5c"} Jan 06 15:15:52 crc kubenswrapper[4744]: I0106 15:15:52.947773 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" podStartSLOduration=2.483521971 podStartE2EDuration="2.947753281s" podCreationTimestamp="2026-01-06 15:15:50 +0000 UTC" firstStartedPulling="2026-01-06 15:15:51.935904985 +0000 UTC m=+2348.563371313" lastFinishedPulling="2026-01-06 15:15:52.400136275 +0000 UTC m=+2349.027602623" observedRunningTime="2026-01-06 15:15:52.93828404 +0000 UTC m=+2349.565750368" watchObservedRunningTime="2026-01-06 15:15:52.947753281 +0000 UTC m=+2349.575219609" Jan 06 15:16:05 crc kubenswrapper[4744]: I0106 15:16:05.712967 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:16:05 crc kubenswrapper[4744]: E0106 15:16:05.714925 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:16:08 crc kubenswrapper[4744]: I0106 15:16:08.049082 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-jndqv"] Jan 06 15:16:08 crc kubenswrapper[4744]: I0106 15:16:08.061550 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-jndqv"] Jan 06 15:16:09 crc kubenswrapper[4744]: I0106 15:16:09.725914 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df07cf75-bdf5-43e2-8066-e1b4f2d7b583" path="/var/lib/kubelet/pods/df07cf75-bdf5-43e2-8066-e1b4f2d7b583/volumes" Jan 06 15:16:11 crc kubenswrapper[4744]: I0106 15:16:11.963355 4744 scope.go:117] "RemoveContainer" containerID="0f07881fd6f2cf7fad444b13357e5f271c058d4fb714372c0b5007681855a1be" Jan 06 15:16:18 crc kubenswrapper[4744]: I0106 15:16:18.711304 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:16:18 crc kubenswrapper[4744]: E0106 15:16:18.712344 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:16:32 crc kubenswrapper[4744]: I0106 15:16:32.712635 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:16:32 crc kubenswrapper[4744]: E0106 15:16:32.713867 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:16:47 crc kubenswrapper[4744]: I0106 15:16:47.711656 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:16:47 crc kubenswrapper[4744]: E0106 15:16:47.712866 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:16:50 crc kubenswrapper[4744]: I0106 15:16:50.645643 4744 generic.go:334] "Generic (PLEG): container finished" podID="2b9ab759-a383-4ec8-ad5a-578f35de2a5e" containerID="3b7ce69223c3dad14ccf3dec41edb75d5ebbfa653c57fa7d5d7ef0311fb0f75c" exitCode=0 Jan 06 15:16:50 crc kubenswrapper[4744]: I0106 15:16:50.645713 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" event={"ID":"2b9ab759-a383-4ec8-ad5a-578f35de2a5e","Type":"ContainerDied","Data":"3b7ce69223c3dad14ccf3dec41edb75d5ebbfa653c57fa7d5d7ef0311fb0f75c"} Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.246094 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.272778 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-inventory\") pod \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.273138 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpp7d\" (UniqueName: \"kubernetes.io/projected/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-kube-api-access-dpp7d\") pod \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.273417 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-ssh-key-openstack-edpm-ipam\") pod \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\" (UID: \"2b9ab759-a383-4ec8-ad5a-578f35de2a5e\") " Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.278828 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-kube-api-access-dpp7d" (OuterVolumeSpecName: "kube-api-access-dpp7d") pod "2b9ab759-a383-4ec8-ad5a-578f35de2a5e" (UID: "2b9ab759-a383-4ec8-ad5a-578f35de2a5e"). InnerVolumeSpecName "kube-api-access-dpp7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.305993 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-inventory" (OuterVolumeSpecName: "inventory") pod "2b9ab759-a383-4ec8-ad5a-578f35de2a5e" (UID: "2b9ab759-a383-4ec8-ad5a-578f35de2a5e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.324880 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2b9ab759-a383-4ec8-ad5a-578f35de2a5e" (UID: "2b9ab759-a383-4ec8-ad5a-578f35de2a5e"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.376746 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.376784 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpp7d\" (UniqueName: \"kubernetes.io/projected/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-kube-api-access-dpp7d\") on node \"crc\" DevicePath \"\"" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.376795 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2b9ab759-a383-4ec8-ad5a-578f35de2a5e-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.681031 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" event={"ID":"2b9ab759-a383-4ec8-ad5a-578f35de2a5e","Type":"ContainerDied","Data":"ca3de53e95b29b7a12cde9c5591f2fb2dc667c49d9fda14e058cb802db67da5c"} Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.681212 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.681240 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca3de53e95b29b7a12cde9c5591f2fb2dc667c49d9fda14e058cb802db67da5c" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.807257 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-km9j8"] Jan 06 15:16:52 crc kubenswrapper[4744]: E0106 15:16:52.808246 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b9ab759-a383-4ec8-ad5a-578f35de2a5e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.808289 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b9ab759-a383-4ec8-ad5a-578f35de2a5e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.808847 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b9ab759-a383-4ec8-ad5a-578f35de2a5e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.810772 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.813992 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.814039 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.814582 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.819473 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-km9j8"] Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.821098 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.889733 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-km9j8\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.890180 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-km9j8\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.890293 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kmbr\" (UniqueName: \"kubernetes.io/projected/e77857ea-77e9-4d71-80bb-01107b6b3820-kube-api-access-7kmbr\") pod \"ssh-known-hosts-edpm-deployment-km9j8\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.992740 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-km9j8\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.992838 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-km9j8\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.992936 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kmbr\" (UniqueName: \"kubernetes.io/projected/e77857ea-77e9-4d71-80bb-01107b6b3820-kube-api-access-7kmbr\") pod \"ssh-known-hosts-edpm-deployment-km9j8\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.997228 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-km9j8\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:52 crc kubenswrapper[4744]: I0106 15:16:52.997601 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-km9j8\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:53 crc kubenswrapper[4744]: I0106 15:16:53.014646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kmbr\" (UniqueName: \"kubernetes.io/projected/e77857ea-77e9-4d71-80bb-01107b6b3820-kube-api-access-7kmbr\") pod \"ssh-known-hosts-edpm-deployment-km9j8\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:53 crc kubenswrapper[4744]: I0106 15:16:53.150793 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:16:53 crc kubenswrapper[4744]: I0106 15:16:53.768179 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-km9j8"] Jan 06 15:16:54 crc kubenswrapper[4744]: I0106 15:16:54.705040 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" event={"ID":"e77857ea-77e9-4d71-80bb-01107b6b3820","Type":"ContainerStarted","Data":"0376716b0d8ccceb44d157e56213548db38bb7c8bba55dbede1ce9f6e57778cc"} Jan 06 15:16:54 crc kubenswrapper[4744]: I0106 15:16:54.705376 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" event={"ID":"e77857ea-77e9-4d71-80bb-01107b6b3820","Type":"ContainerStarted","Data":"8874b77ac9bc1b4e1d8c808f591f7574dce2bf10dc32572302f0c33f6f7b70d8"} Jan 06 15:16:54 crc kubenswrapper[4744]: I0106 15:16:54.745555 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" podStartSLOduration=2.3397249909999998 podStartE2EDuration="2.745525809s" podCreationTimestamp="2026-01-06 15:16:52 +0000 UTC" firstStartedPulling="2026-01-06 15:16:53.769446641 +0000 UTC m=+2410.396912959" lastFinishedPulling="2026-01-06 15:16:54.175247459 +0000 UTC m=+2410.802713777" observedRunningTime="2026-01-06 15:16:54.721198603 +0000 UTC m=+2411.348664931" watchObservedRunningTime="2026-01-06 15:16:54.745525809 +0000 UTC m=+2411.372992167" Jan 06 15:16:59 crc kubenswrapper[4744]: I0106 15:16:59.712573 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:16:59 crc kubenswrapper[4744]: E0106 15:16:59.713645 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:17:02 crc kubenswrapper[4744]: I0106 15:17:02.825486 4744 generic.go:334] "Generic (PLEG): container finished" podID="e77857ea-77e9-4d71-80bb-01107b6b3820" containerID="0376716b0d8ccceb44d157e56213548db38bb7c8bba55dbede1ce9f6e57778cc" exitCode=0 Jan 06 15:17:02 crc kubenswrapper[4744]: I0106 15:17:02.825600 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" event={"ID":"e77857ea-77e9-4d71-80bb-01107b6b3820","Type":"ContainerDied","Data":"0376716b0d8ccceb44d157e56213548db38bb7c8bba55dbede1ce9f6e57778cc"} Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.328221 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.423737 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-inventory-0\") pod \"e77857ea-77e9-4d71-80bb-01107b6b3820\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.424339 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-ssh-key-openstack-edpm-ipam\") pod \"e77857ea-77e9-4d71-80bb-01107b6b3820\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.424643 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kmbr\" (UniqueName: \"kubernetes.io/projected/e77857ea-77e9-4d71-80bb-01107b6b3820-kube-api-access-7kmbr\") pod \"e77857ea-77e9-4d71-80bb-01107b6b3820\" (UID: \"e77857ea-77e9-4d71-80bb-01107b6b3820\") " Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.429512 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e77857ea-77e9-4d71-80bb-01107b6b3820-kube-api-access-7kmbr" (OuterVolumeSpecName: "kube-api-access-7kmbr") pod "e77857ea-77e9-4d71-80bb-01107b6b3820" (UID: "e77857ea-77e9-4d71-80bb-01107b6b3820"). InnerVolumeSpecName "kube-api-access-7kmbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.457260 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e77857ea-77e9-4d71-80bb-01107b6b3820" (UID: "e77857ea-77e9-4d71-80bb-01107b6b3820"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.463570 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e77857ea-77e9-4d71-80bb-01107b6b3820" (UID: "e77857ea-77e9-4d71-80bb-01107b6b3820"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.529760 4744 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-inventory-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.529820 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e77857ea-77e9-4d71-80bb-01107b6b3820-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.529840 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kmbr\" (UniqueName: \"kubernetes.io/projected/e77857ea-77e9-4d71-80bb-01107b6b3820-kube-api-access-7kmbr\") on node \"crc\" DevicePath \"\"" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.857282 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" event={"ID":"e77857ea-77e9-4d71-80bb-01107b6b3820","Type":"ContainerDied","Data":"8874b77ac9bc1b4e1d8c808f591f7574dce2bf10dc32572302f0c33f6f7b70d8"} Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.857330 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8874b77ac9bc1b4e1d8c808f591f7574dce2bf10dc32572302f0c33f6f7b70d8" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.857396 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-km9j8" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.926013 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b"] Jan 06 15:17:04 crc kubenswrapper[4744]: E0106 15:17:04.926738 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e77857ea-77e9-4d71-80bb-01107b6b3820" containerName="ssh-known-hosts-edpm-deployment" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.926827 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e77857ea-77e9-4d71-80bb-01107b6b3820" containerName="ssh-known-hosts-edpm-deployment" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.927233 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e77857ea-77e9-4d71-80bb-01107b6b3820" containerName="ssh-known-hosts-edpm-deployment" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.928190 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.931867 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.932261 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.932341 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.936685 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:17:04 crc kubenswrapper[4744]: I0106 15:17:04.946469 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b"] Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.042467 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-clw5b\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.042605 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqx4z\" (UniqueName: \"kubernetes.io/projected/95e3d07e-c41f-4d70-8890-6b87c90468b0-kube-api-access-sqx4z\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-clw5b\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.042731 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-clw5b\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.145396 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-clw5b\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.145798 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-clw5b\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.146033 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqx4z\" (UniqueName: \"kubernetes.io/projected/95e3d07e-c41f-4d70-8890-6b87c90468b0-kube-api-access-sqx4z\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-clw5b\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.149922 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-clw5b\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.150735 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-clw5b\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.165134 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqx4z\" (UniqueName: \"kubernetes.io/projected/95e3d07e-c41f-4d70-8890-6b87c90468b0-kube-api-access-sqx4z\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-clw5b\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.262835 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:05 crc kubenswrapper[4744]: I0106 15:17:05.673119 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b"] Jan 06 15:17:06 crc kubenswrapper[4744]: I0106 15:17:05.868944 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" event={"ID":"95e3d07e-c41f-4d70-8890-6b87c90468b0","Type":"ContainerStarted","Data":"548a7f34d1778b17b72ea151f5298431c0b046f133f82f2f9061e319fdce2c6b"} Jan 06 15:17:06 crc kubenswrapper[4744]: I0106 15:17:06.889611 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" event={"ID":"95e3d07e-c41f-4d70-8890-6b87c90468b0","Type":"ContainerStarted","Data":"95eb4a479813d604c726edaf9ecdbca7704ad1bf64761892d02b8ad293424fe1"} Jan 06 15:17:06 crc kubenswrapper[4744]: I0106 15:17:06.917206 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" podStartSLOduration=2.3688886399999998 podStartE2EDuration="2.917181866s" podCreationTimestamp="2026-01-06 15:17:04 +0000 UTC" firstStartedPulling="2026-01-06 15:17:05.679207636 +0000 UTC m=+2422.306673954" lastFinishedPulling="2026-01-06 15:17:06.227500852 +0000 UTC m=+2422.854967180" observedRunningTime="2026-01-06 15:17:06.912092481 +0000 UTC m=+2423.539558819" watchObservedRunningTime="2026-01-06 15:17:06.917181866 +0000 UTC m=+2423.544648184" Jan 06 15:17:13 crc kubenswrapper[4744]: I0106 15:17:13.711732 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:17:13 crc kubenswrapper[4744]: E0106 15:17:13.713014 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:17:16 crc kubenswrapper[4744]: I0106 15:17:16.006048 4744 generic.go:334] "Generic (PLEG): container finished" podID="95e3d07e-c41f-4d70-8890-6b87c90468b0" containerID="95eb4a479813d604c726edaf9ecdbca7704ad1bf64761892d02b8ad293424fe1" exitCode=0 Jan 06 15:17:16 crc kubenswrapper[4744]: I0106 15:17:16.006147 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" event={"ID":"95e3d07e-c41f-4d70-8890-6b87c90468b0","Type":"ContainerDied","Data":"95eb4a479813d604c726edaf9ecdbca7704ad1bf64761892d02b8ad293424fe1"} Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.531942 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.711233 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqx4z\" (UniqueName: \"kubernetes.io/projected/95e3d07e-c41f-4d70-8890-6b87c90468b0-kube-api-access-sqx4z\") pod \"95e3d07e-c41f-4d70-8890-6b87c90468b0\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.711302 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-ssh-key-openstack-edpm-ipam\") pod \"95e3d07e-c41f-4d70-8890-6b87c90468b0\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.711390 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-inventory\") pod \"95e3d07e-c41f-4d70-8890-6b87c90468b0\" (UID: \"95e3d07e-c41f-4d70-8890-6b87c90468b0\") " Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.718203 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e3d07e-c41f-4d70-8890-6b87c90468b0-kube-api-access-sqx4z" (OuterVolumeSpecName: "kube-api-access-sqx4z") pod "95e3d07e-c41f-4d70-8890-6b87c90468b0" (UID: "95e3d07e-c41f-4d70-8890-6b87c90468b0"). InnerVolumeSpecName "kube-api-access-sqx4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.746130 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-inventory" (OuterVolumeSpecName: "inventory") pod "95e3d07e-c41f-4d70-8890-6b87c90468b0" (UID: "95e3d07e-c41f-4d70-8890-6b87c90468b0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.761411 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "95e3d07e-c41f-4d70-8890-6b87c90468b0" (UID: "95e3d07e-c41f-4d70-8890-6b87c90468b0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.814582 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqx4z\" (UniqueName: \"kubernetes.io/projected/95e3d07e-c41f-4d70-8890-6b87c90468b0-kube-api-access-sqx4z\") on node \"crc\" DevicePath \"\"" Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.814831 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:17:17 crc kubenswrapper[4744]: I0106 15:17:17.814846 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/95e3d07e-c41f-4d70-8890-6b87c90468b0-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.027306 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" event={"ID":"95e3d07e-c41f-4d70-8890-6b87c90468b0","Type":"ContainerDied","Data":"548a7f34d1778b17b72ea151f5298431c0b046f133f82f2f9061e319fdce2c6b"} Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.027341 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="548a7f34d1778b17b72ea151f5298431c0b046f133f82f2f9061e319fdce2c6b" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.027393 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-clw5b" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.142639 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm"] Jan 06 15:17:18 crc kubenswrapper[4744]: E0106 15:17:18.143124 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e3d07e-c41f-4d70-8890-6b87c90468b0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.143143 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e3d07e-c41f-4d70-8890-6b87c90468b0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.143468 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e3d07e-c41f-4d70-8890-6b87c90468b0" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.146255 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.150405 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.151990 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.152370 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.154310 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5wcb\" (UniqueName: \"kubernetes.io/projected/7bb6bb0d-d95e-48e5-96a0-907532d387d2-kube-api-access-l5wcb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.154564 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.154679 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.155868 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.163077 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm"] Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.256943 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.257019 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.257202 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5wcb\" (UniqueName: \"kubernetes.io/projected/7bb6bb0d-d95e-48e5-96a0-907532d387d2-kube-api-access-l5wcb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.262478 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.262672 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.277952 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5wcb\" (UniqueName: \"kubernetes.io/projected/7bb6bb0d-d95e-48e5-96a0-907532d387d2-kube-api-access-l5wcb\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:18 crc kubenswrapper[4744]: I0106 15:17:18.466205 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:19 crc kubenswrapper[4744]: I0106 15:17:19.081239 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm"] Jan 06 15:17:20 crc kubenswrapper[4744]: I0106 15:17:20.052992 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" event={"ID":"7bb6bb0d-d95e-48e5-96a0-907532d387d2","Type":"ContainerStarted","Data":"ba35d1fce4f744f5955c949819b01cf187a4c1b78fd6bbeed3cacb64f43e5ad2"} Jan 06 15:17:20 crc kubenswrapper[4744]: I0106 15:17:20.053247 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" event={"ID":"7bb6bb0d-d95e-48e5-96a0-907532d387d2","Type":"ContainerStarted","Data":"3534aa974b23d0e4e2f4e0b250b4ba7ccf334876381df0d7b72ffcf534c25d4c"} Jan 06 15:17:20 crc kubenswrapper[4744]: I0106 15:17:20.076179 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" podStartSLOduration=1.520433326 podStartE2EDuration="2.076123848s" podCreationTimestamp="2026-01-06 15:17:18 +0000 UTC" firstStartedPulling="2026-01-06 15:17:19.091823082 +0000 UTC m=+2435.719289400" lastFinishedPulling="2026-01-06 15:17:19.647513604 +0000 UTC m=+2436.274979922" observedRunningTime="2026-01-06 15:17:20.072551923 +0000 UTC m=+2436.700018251" watchObservedRunningTime="2026-01-06 15:17:20.076123848 +0000 UTC m=+2436.703590196" Jan 06 15:17:27 crc kubenswrapper[4744]: I0106 15:17:27.711682 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:17:27 crc kubenswrapper[4744]: E0106 15:17:27.712599 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:17:31 crc kubenswrapper[4744]: I0106 15:17:31.185802 4744 generic.go:334] "Generic (PLEG): container finished" podID="7bb6bb0d-d95e-48e5-96a0-907532d387d2" containerID="ba35d1fce4f744f5955c949819b01cf187a4c1b78fd6bbeed3cacb64f43e5ad2" exitCode=0 Jan 06 15:17:31 crc kubenswrapper[4744]: I0106 15:17:31.185903 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" event={"ID":"7bb6bb0d-d95e-48e5-96a0-907532d387d2","Type":"ContainerDied","Data":"ba35d1fce4f744f5955c949819b01cf187a4c1b78fd6bbeed3cacb64f43e5ad2"} Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.642286 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.837977 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-ssh-key-openstack-edpm-ipam\") pod \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.838048 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5wcb\" (UniqueName: \"kubernetes.io/projected/7bb6bb0d-d95e-48e5-96a0-907532d387d2-kube-api-access-l5wcb\") pod \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.838362 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-inventory\") pod \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\" (UID: \"7bb6bb0d-d95e-48e5-96a0-907532d387d2\") " Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.857438 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb6bb0d-d95e-48e5-96a0-907532d387d2-kube-api-access-l5wcb" (OuterVolumeSpecName: "kube-api-access-l5wcb") pod "7bb6bb0d-d95e-48e5-96a0-907532d387d2" (UID: "7bb6bb0d-d95e-48e5-96a0-907532d387d2"). InnerVolumeSpecName "kube-api-access-l5wcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.897972 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7bb6bb0d-d95e-48e5-96a0-907532d387d2" (UID: "7bb6bb0d-d95e-48e5-96a0-907532d387d2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.899004 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-inventory" (OuterVolumeSpecName: "inventory") pod "7bb6bb0d-d95e-48e5-96a0-907532d387d2" (UID: "7bb6bb0d-d95e-48e5-96a0-907532d387d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.941692 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.941732 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5wcb\" (UniqueName: \"kubernetes.io/projected/7bb6bb0d-d95e-48e5-96a0-907532d387d2-kube-api-access-l5wcb\") on node \"crc\" DevicePath \"\"" Jan 06 15:17:32 crc kubenswrapper[4744]: I0106 15:17:32.941743 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7bb6bb0d-d95e-48e5-96a0-907532d387d2-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.216281 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" event={"ID":"7bb6bb0d-d95e-48e5-96a0-907532d387d2","Type":"ContainerDied","Data":"3534aa974b23d0e4e2f4e0b250b4ba7ccf334876381df0d7b72ffcf534c25d4c"} Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.216344 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3534aa974b23d0e4e2f4e0b250b4ba7ccf334876381df0d7b72ffcf534c25d4c" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.216384 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.328759 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz"] Jan 06 15:17:33 crc kubenswrapper[4744]: E0106 15:17:33.329401 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bb6bb0d-d95e-48e5-96a0-907532d387d2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.329425 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bb6bb0d-d95e-48e5-96a0-907532d387d2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.329725 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bb6bb0d-d95e-48e5-96a0-907532d387d2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.330817 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.333739 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.333986 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.334242 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.334377 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.334689 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.334745 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.334701 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.336260 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.336680 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.362367 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz"] Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.453915 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.453985 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454124 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454226 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454288 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454392 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454462 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454497 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454596 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454735 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454785 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhj9q\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-kube-api-access-xhj9q\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454832 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454905 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454937 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.454979 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.455303 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558092 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558261 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558404 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558574 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558642 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhj9q\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-kube-api-access-xhj9q\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558735 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558847 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558906 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.558982 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.559207 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.559328 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.559393 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.559470 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.559530 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.559588 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.564556 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.566143 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.566325 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.567227 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.568136 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.568827 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.569600 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.569719 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.569875 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.570139 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.575946 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.577556 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.578678 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.579284 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.581620 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhj9q\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-kube-api-access-xhj9q\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.599248 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:33 crc kubenswrapper[4744]: I0106 15:17:33.686027 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:17:34 crc kubenswrapper[4744]: W0106 15:17:34.341759 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f9705e7_7499_4378_8d58_c9373c83e5ed.slice/crio-fcd55e7be821bc9515e3d8dde2e1a4a9f7312fd3e0412527fefdba9a6d315299 WatchSource:0}: Error finding container fcd55e7be821bc9515e3d8dde2e1a4a9f7312fd3e0412527fefdba9a6d315299: Status 404 returned error can't find the container with id fcd55e7be821bc9515e3d8dde2e1a4a9f7312fd3e0412527fefdba9a6d315299 Jan 06 15:17:34 crc kubenswrapper[4744]: I0106 15:17:34.346505 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz"] Jan 06 15:17:35 crc kubenswrapper[4744]: I0106 15:17:35.241744 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" event={"ID":"5f9705e7-7499-4378-8d58-c9373c83e5ed","Type":"ContainerStarted","Data":"1ab5a6db097e5acd728008a685308466ab0256534a41a62f0c85951b4d829630"} Jan 06 15:17:35 crc kubenswrapper[4744]: I0106 15:17:35.242064 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" event={"ID":"5f9705e7-7499-4378-8d58-c9373c83e5ed","Type":"ContainerStarted","Data":"fcd55e7be821bc9515e3d8dde2e1a4a9f7312fd3e0412527fefdba9a6d315299"} Jan 06 15:17:35 crc kubenswrapper[4744]: I0106 15:17:35.270527 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" podStartSLOduration=1.8080994970000002 podStartE2EDuration="2.27050169s" podCreationTimestamp="2026-01-06 15:17:33 +0000 UTC" firstStartedPulling="2026-01-06 15:17:34.345030407 +0000 UTC m=+2450.972496755" lastFinishedPulling="2026-01-06 15:17:34.80743262 +0000 UTC m=+2451.434898948" observedRunningTime="2026-01-06 15:17:35.263480423 +0000 UTC m=+2451.890946751" watchObservedRunningTime="2026-01-06 15:17:35.27050169 +0000 UTC m=+2451.897968018" Jan 06 15:17:38 crc kubenswrapper[4744]: I0106 15:17:38.711840 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:17:38 crc kubenswrapper[4744]: E0106 15:17:38.713259 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:17:52 crc kubenswrapper[4744]: I0106 15:17:52.711892 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:17:52 crc kubenswrapper[4744]: E0106 15:17:52.713482 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:18:06 crc kubenswrapper[4744]: I0106 15:18:06.714335 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:18:06 crc kubenswrapper[4744]: E0106 15:18:06.716519 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:18:21 crc kubenswrapper[4744]: I0106 15:18:21.711113 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:18:22 crc kubenswrapper[4744]: I0106 15:18:22.925340 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"be2a215fc7aa9e274d639fb1b33f36c687f7a1d28458e250954de94d41c49b35"} Jan 06 15:18:26 crc kubenswrapper[4744]: I0106 15:18:26.970291 4744 generic.go:334] "Generic (PLEG): container finished" podID="5f9705e7-7499-4378-8d58-c9373c83e5ed" containerID="1ab5a6db097e5acd728008a685308466ab0256534a41a62f0c85951b4d829630" exitCode=0 Jan 06 15:18:26 crc kubenswrapper[4744]: I0106 15:18:26.970407 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" event={"ID":"5f9705e7-7499-4378-8d58-c9373c83e5ed","Type":"ContainerDied","Data":"1ab5a6db097e5acd728008a685308466ab0256534a41a62f0c85951b4d829630"} Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.495264 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.580757 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-repo-setup-combined-ca-bundle\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581069 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-ovn-default-certs-0\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581099 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581316 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581369 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-combined-ca-bundle\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581393 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-power-monitoring-combined-ca-bundle\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581416 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-neutron-metadata-combined-ca-bundle\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581445 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-nova-combined-ca-bundle\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581459 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-libvirt-combined-ca-bundle\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581482 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhj9q\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-kube-api-access-xhj9q\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581497 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581582 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ovn-combined-ca-bundle\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581612 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-inventory\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581651 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-bootstrap-combined-ca-bundle\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581696 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.581733 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ssh-key-openstack-edpm-ipam\") pod \"5f9705e7-7499-4378-8d58-c9373c83e5ed\" (UID: \"5f9705e7-7499-4378-8d58-c9373c83e5ed\") " Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.589380 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.589458 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.590320 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.591757 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.591756 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.592568 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.592717 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.593765 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.593812 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.594005 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-kube-api-access-xhj9q" (OuterVolumeSpecName: "kube-api-access-xhj9q") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "kube-api-access-xhj9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.594081 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.594941 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.595344 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.596293 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.620208 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.626835 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-inventory" (OuterVolumeSpecName: "inventory") pod "5f9705e7-7499-4378-8d58-c9373c83e5ed" (UID: "5f9705e7-7499-4378-8d58-c9373c83e5ed"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.684921 4744 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.684959 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.684972 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.684987 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685003 4744 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685016 4744 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685027 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685038 4744 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685047 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685057 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhj9q\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-kube-api-access-xhj9q\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685068 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685081 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685091 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685101 4744 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685113 4744 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5f9705e7-7499-4378-8d58-c9373c83e5ed-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.685124 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5f9705e7-7499-4378-8d58-c9373c83e5ed-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.994873 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" event={"ID":"5f9705e7-7499-4378-8d58-c9373c83e5ed","Type":"ContainerDied","Data":"fcd55e7be821bc9515e3d8dde2e1a4a9f7312fd3e0412527fefdba9a6d315299"} Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.994913 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcd55e7be821bc9515e3d8dde2e1a4a9f7312fd3e0412527fefdba9a6d315299" Jan 06 15:18:28 crc kubenswrapper[4744]: I0106 15:18:28.994963 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.127264 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f"] Jan 06 15:18:29 crc kubenswrapper[4744]: E0106 15:18:29.127698 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f9705e7-7499-4378-8d58-c9373c83e5ed" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.127715 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f9705e7-7499-4378-8d58-c9373c83e5ed" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.127919 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f9705e7-7499-4378-8d58-c9373c83e5ed" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.128631 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.132385 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.132607 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.132922 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.132968 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.137207 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.146520 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f"] Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.301456 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.301693 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.301870 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.302203 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.302407 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4smc\" (UniqueName: \"kubernetes.io/projected/fcdb07a4-62d3-4425-9319-4e58bcb6831c-kube-api-access-x4smc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.405312 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.405394 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.405444 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.405537 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.405588 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4smc\" (UniqueName: \"kubernetes.io/projected/fcdb07a4-62d3-4425-9319-4e58bcb6831c-kube-api-access-x4smc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.406297 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.409473 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.409731 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.416072 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.421857 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4smc\" (UniqueName: \"kubernetes.io/projected/fcdb07a4-62d3-4425-9319-4e58bcb6831c-kube-api-access-x4smc\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-twr7f\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:29 crc kubenswrapper[4744]: I0106 15:18:29.526514 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:18:30 crc kubenswrapper[4744]: I0106 15:18:30.142627 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f"] Jan 06 15:18:31 crc kubenswrapper[4744]: I0106 15:18:31.022044 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" event={"ID":"fcdb07a4-62d3-4425-9319-4e58bcb6831c","Type":"ContainerStarted","Data":"af1e33b19fe08516b4fbd527cbe4ba6df666763cddff850a9902d33f23c9066b"} Jan 06 15:18:31 crc kubenswrapper[4744]: I0106 15:18:31.022371 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" event={"ID":"fcdb07a4-62d3-4425-9319-4e58bcb6831c","Type":"ContainerStarted","Data":"8f0c6e53e277322207f5dcb09b7a3cb4f937b7fdb3a9f855eebbaaf0bd62cc0e"} Jan 06 15:18:31 crc kubenswrapper[4744]: I0106 15:18:31.049429 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" podStartSLOduration=1.5193919230000001 podStartE2EDuration="2.049399583s" podCreationTimestamp="2026-01-06 15:18:29 +0000 UTC" firstStartedPulling="2026-01-06 15:18:30.150763814 +0000 UTC m=+2506.778230142" lastFinishedPulling="2026-01-06 15:18:30.680771484 +0000 UTC m=+2507.308237802" observedRunningTime="2026-01-06 15:18:31.041860793 +0000 UTC m=+2507.669327101" watchObservedRunningTime="2026-01-06 15:18:31.049399583 +0000 UTC m=+2507.676865941" Jan 06 15:19:47 crc kubenswrapper[4744]: I0106 15:19:47.998803 4744 generic.go:334] "Generic (PLEG): container finished" podID="fcdb07a4-62d3-4425-9319-4e58bcb6831c" containerID="af1e33b19fe08516b4fbd527cbe4ba6df666763cddff850a9902d33f23c9066b" exitCode=0 Jan 06 15:19:47 crc kubenswrapper[4744]: I0106 15:19:47.998881 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" event={"ID":"fcdb07a4-62d3-4425-9319-4e58bcb6831c","Type":"ContainerDied","Data":"af1e33b19fe08516b4fbd527cbe4ba6df666763cddff850a9902d33f23c9066b"} Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.599075 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.687558 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4smc\" (UniqueName: \"kubernetes.io/projected/fcdb07a4-62d3-4425-9319-4e58bcb6831c-kube-api-access-x4smc\") pod \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.687661 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-inventory\") pod \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.687691 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovn-combined-ca-bundle\") pod \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.687823 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovncontroller-config-0\") pod \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.687948 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ssh-key-openstack-edpm-ipam\") pod \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\" (UID: \"fcdb07a4-62d3-4425-9319-4e58bcb6831c\") " Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.693011 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "fcdb07a4-62d3-4425-9319-4e58bcb6831c" (UID: "fcdb07a4-62d3-4425-9319-4e58bcb6831c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.699055 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcdb07a4-62d3-4425-9319-4e58bcb6831c-kube-api-access-x4smc" (OuterVolumeSpecName: "kube-api-access-x4smc") pod "fcdb07a4-62d3-4425-9319-4e58bcb6831c" (UID: "fcdb07a4-62d3-4425-9319-4e58bcb6831c"). InnerVolumeSpecName "kube-api-access-x4smc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.730072 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "fcdb07a4-62d3-4425-9319-4e58bcb6831c" (UID: "fcdb07a4-62d3-4425-9319-4e58bcb6831c"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.734705 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "fcdb07a4-62d3-4425-9319-4e58bcb6831c" (UID: "fcdb07a4-62d3-4425-9319-4e58bcb6831c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.735314 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-inventory" (OuterVolumeSpecName: "inventory") pod "fcdb07a4-62d3-4425-9319-4e58bcb6831c" (UID: "fcdb07a4-62d3-4425-9319-4e58bcb6831c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.791047 4744 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.791075 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.791086 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4smc\" (UniqueName: \"kubernetes.io/projected/fcdb07a4-62d3-4425-9319-4e58bcb6831c-kube-api-access-x4smc\") on node \"crc\" DevicePath \"\"" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.791120 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:19:49 crc kubenswrapper[4744]: I0106 15:19:49.791130 4744 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcdb07a4-62d3-4425-9319-4e58bcb6831c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.021482 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" event={"ID":"fcdb07a4-62d3-4425-9319-4e58bcb6831c","Type":"ContainerDied","Data":"8f0c6e53e277322207f5dcb09b7a3cb4f937b7fdb3a9f855eebbaaf0bd62cc0e"} Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.021861 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f0c6e53e277322207f5dcb09b7a3cb4f937b7fdb3a9f855eebbaaf0bd62cc0e" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.021589 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-twr7f" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.230423 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j"] Jan 06 15:19:50 crc kubenswrapper[4744]: E0106 15:19:50.231021 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcdb07a4-62d3-4425-9319-4e58bcb6831c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.231046 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcdb07a4-62d3-4425-9319-4e58bcb6831c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.231497 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcdb07a4-62d3-4425-9319-4e58bcb6831c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.232897 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.235748 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.236771 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.237107 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.237338 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.237785 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.241151 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j"] Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.250917 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.302334 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.302370 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.302406 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.302578 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.303094 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.303127 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psrt2\" (UniqueName: \"kubernetes.io/projected/c913d43d-9615-4436-ae90-56c68245d4a6-kube-api-access-psrt2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.405197 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.405258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psrt2\" (UniqueName: \"kubernetes.io/projected/c913d43d-9615-4436-ae90-56c68245d4a6-kube-api-access-psrt2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.405344 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.405370 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.405415 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.405480 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.410939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.413864 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.415829 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.425647 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.434533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.473070 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psrt2\" (UniqueName: \"kubernetes.io/projected/c913d43d-9615-4436-ae90-56c68245d4a6-kube-api-access-psrt2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:50 crc kubenswrapper[4744]: I0106 15:19:50.556972 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:19:51 crc kubenswrapper[4744]: I0106 15:19:51.141478 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j"] Jan 06 15:19:52 crc kubenswrapper[4744]: I0106 15:19:52.057244 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" event={"ID":"c913d43d-9615-4436-ae90-56c68245d4a6","Type":"ContainerStarted","Data":"a90505b7ed2e3df50fe100275bb67208e4ee4f6c61dc7829fe25dec72dacfa4f"} Jan 06 15:19:52 crc kubenswrapper[4744]: I0106 15:19:52.057596 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" event={"ID":"c913d43d-9615-4436-ae90-56c68245d4a6","Type":"ContainerStarted","Data":"1763b73e61111243b3311e2d33d71752bcd2aa9aab815641f287f46148c49ef7"} Jan 06 15:19:52 crc kubenswrapper[4744]: I0106 15:19:52.103720 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" podStartSLOduration=1.624403329 podStartE2EDuration="2.10368974s" podCreationTimestamp="2026-01-06 15:19:50 +0000 UTC" firstStartedPulling="2026-01-06 15:19:51.135989315 +0000 UTC m=+2587.763455633" lastFinishedPulling="2026-01-06 15:19:51.615275676 +0000 UTC m=+2588.242742044" observedRunningTime="2026-01-06 15:19:52.083579705 +0000 UTC m=+2588.711046083" watchObservedRunningTime="2026-01-06 15:19:52.10368974 +0000 UTC m=+2588.731156098" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.436675 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fwk6x"] Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.439611 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.464399 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fwk6x"] Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.579225 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-catalog-content\") pod \"redhat-operators-fwk6x\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.579275 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-utilities\") pod \"redhat-operators-fwk6x\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.579555 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hltnh\" (UniqueName: \"kubernetes.io/projected/d37f7f03-5fa8-4b00-871c-4f43aac4026f-kube-api-access-hltnh\") pod \"redhat-operators-fwk6x\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.682265 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-utilities\") pod \"redhat-operators-fwk6x\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.682516 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hltnh\" (UniqueName: \"kubernetes.io/projected/d37f7f03-5fa8-4b00-871c-4f43aac4026f-kube-api-access-hltnh\") pod \"redhat-operators-fwk6x\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.682624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-catalog-content\") pod \"redhat-operators-fwk6x\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.682861 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-utilities\") pod \"redhat-operators-fwk6x\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.683106 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-catalog-content\") pod \"redhat-operators-fwk6x\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.708859 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hltnh\" (UniqueName: \"kubernetes.io/projected/d37f7f03-5fa8-4b00-871c-4f43aac4026f-kube-api-access-hltnh\") pod \"redhat-operators-fwk6x\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:11 crc kubenswrapper[4744]: I0106 15:20:11.759529 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:12 crc kubenswrapper[4744]: I0106 15:20:12.358553 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fwk6x"] Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.312595 4744 generic.go:334] "Generic (PLEG): container finished" podID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerID="afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8" exitCode=0 Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.312693 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwk6x" event={"ID":"d37f7f03-5fa8-4b00-871c-4f43aac4026f","Type":"ContainerDied","Data":"afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8"} Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.312863 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwk6x" event={"ID":"d37f7f03-5fa8-4b00-871c-4f43aac4026f","Type":"ContainerStarted","Data":"28ae260b77c303cdcd915416047816e2ea1cc5aac73ab196cdfb0f4cb8de31c6"} Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.314680 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.840054 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lzc4b"] Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.842863 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.874152 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lzc4b"] Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.949711 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-utilities\") pod \"certified-operators-lzc4b\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.949770 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-catalog-content\") pod \"certified-operators-lzc4b\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:13 crc kubenswrapper[4744]: I0106 15:20:13.949802 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gx45\" (UniqueName: \"kubernetes.io/projected/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-kube-api-access-6gx45\") pod \"certified-operators-lzc4b\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:14 crc kubenswrapper[4744]: I0106 15:20:14.052665 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-utilities\") pod \"certified-operators-lzc4b\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:14 crc kubenswrapper[4744]: I0106 15:20:14.052757 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-catalog-content\") pod \"certified-operators-lzc4b\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:14 crc kubenswrapper[4744]: I0106 15:20:14.053074 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-utilities\") pod \"certified-operators-lzc4b\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:14 crc kubenswrapper[4744]: I0106 15:20:14.053436 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-catalog-content\") pod \"certified-operators-lzc4b\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:14 crc kubenswrapper[4744]: I0106 15:20:14.053479 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gx45\" (UniqueName: \"kubernetes.io/projected/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-kube-api-access-6gx45\") pod \"certified-operators-lzc4b\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:14 crc kubenswrapper[4744]: I0106 15:20:14.074011 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gx45\" (UniqueName: \"kubernetes.io/projected/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-kube-api-access-6gx45\") pod \"certified-operators-lzc4b\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:14 crc kubenswrapper[4744]: I0106 15:20:14.177967 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:14 crc kubenswrapper[4744]: W0106 15:20:14.710276 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ada8e8c_6c87_441e_8a39_cca5d58b3da0.slice/crio-f057ded4a9356abb58d427a6911326f6e5af20a3fcbab38cfb705171dd94590f WatchSource:0}: Error finding container f057ded4a9356abb58d427a6911326f6e5af20a3fcbab38cfb705171dd94590f: Status 404 returned error can't find the container with id f057ded4a9356abb58d427a6911326f6e5af20a3fcbab38cfb705171dd94590f Jan 06 15:20:14 crc kubenswrapper[4744]: I0106 15:20:14.713635 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lzc4b"] Jan 06 15:20:15 crc kubenswrapper[4744]: I0106 15:20:15.363171 4744 generic.go:334] "Generic (PLEG): container finished" podID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerID="b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3" exitCode=0 Jan 06 15:20:15 crc kubenswrapper[4744]: I0106 15:20:15.363324 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzc4b" event={"ID":"2ada8e8c-6c87-441e-8a39-cca5d58b3da0","Type":"ContainerDied","Data":"b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3"} Jan 06 15:20:15 crc kubenswrapper[4744]: I0106 15:20:15.363500 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzc4b" event={"ID":"2ada8e8c-6c87-441e-8a39-cca5d58b3da0","Type":"ContainerStarted","Data":"f057ded4a9356abb58d427a6911326f6e5af20a3fcbab38cfb705171dd94590f"} Jan 06 15:20:15 crc kubenswrapper[4744]: I0106 15:20:15.365294 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwk6x" event={"ID":"d37f7f03-5fa8-4b00-871c-4f43aac4026f","Type":"ContainerStarted","Data":"cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9"} Jan 06 15:20:17 crc kubenswrapper[4744]: I0106 15:20:17.393807 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzc4b" event={"ID":"2ada8e8c-6c87-441e-8a39-cca5d58b3da0","Type":"ContainerStarted","Data":"8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22"} Jan 06 15:20:19 crc kubenswrapper[4744]: I0106 15:20:19.436393 4744 generic.go:334] "Generic (PLEG): container finished" podID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerID="cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9" exitCode=0 Jan 06 15:20:19 crc kubenswrapper[4744]: I0106 15:20:19.436456 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwk6x" event={"ID":"d37f7f03-5fa8-4b00-871c-4f43aac4026f","Type":"ContainerDied","Data":"cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9"} Jan 06 15:20:20 crc kubenswrapper[4744]: I0106 15:20:20.458056 4744 generic.go:334] "Generic (PLEG): container finished" podID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerID="8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22" exitCode=0 Jan 06 15:20:20 crc kubenswrapper[4744]: I0106 15:20:20.458129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzc4b" event={"ID":"2ada8e8c-6c87-441e-8a39-cca5d58b3da0","Type":"ContainerDied","Data":"8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22"} Jan 06 15:20:21 crc kubenswrapper[4744]: I0106 15:20:21.471436 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzc4b" event={"ID":"2ada8e8c-6c87-441e-8a39-cca5d58b3da0","Type":"ContainerStarted","Data":"57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68"} Jan 06 15:20:21 crc kubenswrapper[4744]: I0106 15:20:21.474259 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwk6x" event={"ID":"d37f7f03-5fa8-4b00-871c-4f43aac4026f","Type":"ContainerStarted","Data":"544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f"} Jan 06 15:20:21 crc kubenswrapper[4744]: I0106 15:20:21.494633 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lzc4b" podStartSLOduration=2.784946351 podStartE2EDuration="8.494611195s" podCreationTimestamp="2026-01-06 15:20:13 +0000 UTC" firstStartedPulling="2026-01-06 15:20:15.365342064 +0000 UTC m=+2611.992808382" lastFinishedPulling="2026-01-06 15:20:21.075006908 +0000 UTC m=+2617.702473226" observedRunningTime="2026-01-06 15:20:21.493460274 +0000 UTC m=+2618.120926602" watchObservedRunningTime="2026-01-06 15:20:21.494611195 +0000 UTC m=+2618.122077513" Jan 06 15:20:21 crc kubenswrapper[4744]: I0106 15:20:21.529618 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fwk6x" podStartSLOduration=3.576074488 podStartE2EDuration="10.529600545s" podCreationTimestamp="2026-01-06 15:20:11 +0000 UTC" firstStartedPulling="2026-01-06 15:20:13.314493864 +0000 UTC m=+2609.941960172" lastFinishedPulling="2026-01-06 15:20:20.268019911 +0000 UTC m=+2616.895486229" observedRunningTime="2026-01-06 15:20:21.519557028 +0000 UTC m=+2618.147023356" watchObservedRunningTime="2026-01-06 15:20:21.529600545 +0000 UTC m=+2618.157066863" Jan 06 15:20:21 crc kubenswrapper[4744]: I0106 15:20:21.759718 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:21 crc kubenswrapper[4744]: I0106 15:20:21.759806 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:22 crc kubenswrapper[4744]: I0106 15:20:22.818129 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fwk6x" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="registry-server" probeResult="failure" output=< Jan 06 15:20:22 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 15:20:22 crc kubenswrapper[4744]: > Jan 06 15:20:24 crc kubenswrapper[4744]: I0106 15:20:24.180806 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:24 crc kubenswrapper[4744]: I0106 15:20:24.182203 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:24 crc kubenswrapper[4744]: I0106 15:20:24.237337 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:32 crc kubenswrapper[4744]: I0106 15:20:32.830813 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fwk6x" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="registry-server" probeResult="failure" output=< Jan 06 15:20:32 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 15:20:32 crc kubenswrapper[4744]: > Jan 06 15:20:34 crc kubenswrapper[4744]: I0106 15:20:34.230752 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:34 crc kubenswrapper[4744]: I0106 15:20:34.288650 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lzc4b"] Jan 06 15:20:34 crc kubenswrapper[4744]: I0106 15:20:34.642109 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lzc4b" podUID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerName="registry-server" containerID="cri-o://57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68" gracePeriod=2 Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.190028 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.295590 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-utilities\") pod \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.295722 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-catalog-content\") pod \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.295877 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gx45\" (UniqueName: \"kubernetes.io/projected/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-kube-api-access-6gx45\") pod \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\" (UID: \"2ada8e8c-6c87-441e-8a39-cca5d58b3da0\") " Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.296261 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-utilities" (OuterVolumeSpecName: "utilities") pod "2ada8e8c-6c87-441e-8a39-cca5d58b3da0" (UID: "2ada8e8c-6c87-441e-8a39-cca5d58b3da0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.296563 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.302240 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-kube-api-access-6gx45" (OuterVolumeSpecName: "kube-api-access-6gx45") pod "2ada8e8c-6c87-441e-8a39-cca5d58b3da0" (UID: "2ada8e8c-6c87-441e-8a39-cca5d58b3da0"). InnerVolumeSpecName "kube-api-access-6gx45". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.350221 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ada8e8c-6c87-441e-8a39-cca5d58b3da0" (UID: "2ada8e8c-6c87-441e-8a39-cca5d58b3da0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.398793 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gx45\" (UniqueName: \"kubernetes.io/projected/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-kube-api-access-6gx45\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.398824 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ada8e8c-6c87-441e-8a39-cca5d58b3da0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.660029 4744 generic.go:334] "Generic (PLEG): container finished" podID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerID="57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68" exitCode=0 Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.660087 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzc4b" event={"ID":"2ada8e8c-6c87-441e-8a39-cca5d58b3da0","Type":"ContainerDied","Data":"57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68"} Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.660127 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lzc4b" event={"ID":"2ada8e8c-6c87-441e-8a39-cca5d58b3da0","Type":"ContainerDied","Data":"f057ded4a9356abb58d427a6911326f6e5af20a3fcbab38cfb705171dd94590f"} Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.660149 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lzc4b" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.660162 4744 scope.go:117] "RemoveContainer" containerID="57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.687861 4744 scope.go:117] "RemoveContainer" containerID="8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.727563 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lzc4b"] Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.731594 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lzc4b"] Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.732358 4744 scope.go:117] "RemoveContainer" containerID="b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.801844 4744 scope.go:117] "RemoveContainer" containerID="57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68" Jan 06 15:20:35 crc kubenswrapper[4744]: E0106 15:20:35.802257 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68\": container with ID starting with 57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68 not found: ID does not exist" containerID="57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.802314 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68"} err="failed to get container status \"57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68\": rpc error: code = NotFound desc = could not find container \"57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68\": container with ID starting with 57c7e44cec2dc3b821e49df11c52c4527cc4cd15121076ed4270e32a053bcc68 not found: ID does not exist" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.802347 4744 scope.go:117] "RemoveContainer" containerID="8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22" Jan 06 15:20:35 crc kubenswrapper[4744]: E0106 15:20:35.802682 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22\": container with ID starting with 8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22 not found: ID does not exist" containerID="8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.802759 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22"} err="failed to get container status \"8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22\": rpc error: code = NotFound desc = could not find container \"8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22\": container with ID starting with 8026243a6d471101ef34f0f28515ffae528ab94414b5229b005d5c2e3a8daa22 not found: ID does not exist" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.802792 4744 scope.go:117] "RemoveContainer" containerID="b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3" Jan 06 15:20:35 crc kubenswrapper[4744]: E0106 15:20:35.803132 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3\": container with ID starting with b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3 not found: ID does not exist" containerID="b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3" Jan 06 15:20:35 crc kubenswrapper[4744]: I0106 15:20:35.803163 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3"} err="failed to get container status \"b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3\": rpc error: code = NotFound desc = could not find container \"b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3\": container with ID starting with b3dbe0780e15880cef58780dcaa121ea0b89881f45b4a8474c1399acc181aba3 not found: ID does not exist" Jan 06 15:20:37 crc kubenswrapper[4744]: I0106 15:20:37.729420 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" path="/var/lib/kubelet/pods/2ada8e8c-6c87-441e-8a39-cca5d58b3da0/volumes" Jan 06 15:20:41 crc kubenswrapper[4744]: I0106 15:20:41.837509 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:41 crc kubenswrapper[4744]: I0106 15:20:41.920087 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:42 crc kubenswrapper[4744]: I0106 15:20:42.079566 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fwk6x"] Jan 06 15:20:43 crc kubenswrapper[4744]: I0106 15:20:43.777434 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fwk6x" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="registry-server" containerID="cri-o://544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f" gracePeriod=2 Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.394313 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.423774 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.423855 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.449130 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-utilities\") pod \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.449557 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hltnh\" (UniqueName: \"kubernetes.io/projected/d37f7f03-5fa8-4b00-871c-4f43aac4026f-kube-api-access-hltnh\") pod \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.449833 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-catalog-content\") pod \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\" (UID: \"d37f7f03-5fa8-4b00-871c-4f43aac4026f\") " Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.450490 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-utilities" (OuterVolumeSpecName: "utilities") pod "d37f7f03-5fa8-4b00-871c-4f43aac4026f" (UID: "d37f7f03-5fa8-4b00-871c-4f43aac4026f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.451412 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.460212 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d37f7f03-5fa8-4b00-871c-4f43aac4026f-kube-api-access-hltnh" (OuterVolumeSpecName: "kube-api-access-hltnh") pod "d37f7f03-5fa8-4b00-871c-4f43aac4026f" (UID: "d37f7f03-5fa8-4b00-871c-4f43aac4026f"). InnerVolumeSpecName "kube-api-access-hltnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.553502 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hltnh\" (UniqueName: \"kubernetes.io/projected/d37f7f03-5fa8-4b00-871c-4f43aac4026f-kube-api-access-hltnh\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.587816 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d37f7f03-5fa8-4b00-871c-4f43aac4026f" (UID: "d37f7f03-5fa8-4b00-871c-4f43aac4026f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.655980 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d37f7f03-5fa8-4b00-871c-4f43aac4026f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.792416 4744 generic.go:334] "Generic (PLEG): container finished" podID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerID="544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f" exitCode=0 Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.792465 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwk6x" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.792469 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwk6x" event={"ID":"d37f7f03-5fa8-4b00-871c-4f43aac4026f","Type":"ContainerDied","Data":"544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f"} Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.792548 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwk6x" event={"ID":"d37f7f03-5fa8-4b00-871c-4f43aac4026f","Type":"ContainerDied","Data":"28ae260b77c303cdcd915416047816e2ea1cc5aac73ab196cdfb0f4cb8de31c6"} Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.792580 4744 scope.go:117] "RemoveContainer" containerID="544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.812379 4744 scope.go:117] "RemoveContainer" containerID="cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.836155 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fwk6x"] Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.846928 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fwk6x"] Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.847553 4744 scope.go:117] "RemoveContainer" containerID="afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.896468 4744 scope.go:117] "RemoveContainer" containerID="544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f" Jan 06 15:20:44 crc kubenswrapper[4744]: E0106 15:20:44.896850 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f\": container with ID starting with 544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f not found: ID does not exist" containerID="544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.896902 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f"} err="failed to get container status \"544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f\": rpc error: code = NotFound desc = could not find container \"544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f\": container with ID starting with 544f828ade5c1f805209624a7760b99a2c9078ed9fea47cda15011334122cb3f not found: ID does not exist" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.896933 4744 scope.go:117] "RemoveContainer" containerID="cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9" Jan 06 15:20:44 crc kubenswrapper[4744]: E0106 15:20:44.897241 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9\": container with ID starting with cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9 not found: ID does not exist" containerID="cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.897267 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9"} err="failed to get container status \"cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9\": rpc error: code = NotFound desc = could not find container \"cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9\": container with ID starting with cbede90d5ea8617253dc65a114ff51664403c8addcef1c657d4e62d84dbb76d9 not found: ID does not exist" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.897290 4744 scope.go:117] "RemoveContainer" containerID="afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8" Jan 06 15:20:44 crc kubenswrapper[4744]: E0106 15:20:44.897616 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8\": container with ID starting with afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8 not found: ID does not exist" containerID="afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8" Jan 06 15:20:44 crc kubenswrapper[4744]: I0106 15:20:44.897654 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8"} err="failed to get container status \"afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8\": rpc error: code = NotFound desc = could not find container \"afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8\": container with ID starting with afc621057297cbc5e854d4db4c1201079a5d52363dff7b2f8aec7871bed395a8 not found: ID does not exist" Jan 06 15:20:45 crc kubenswrapper[4744]: I0106 15:20:45.750407 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" path="/var/lib/kubelet/pods/d37f7f03-5fa8-4b00-871c-4f43aac4026f/volumes" Jan 06 15:20:55 crc kubenswrapper[4744]: I0106 15:20:55.930378 4744 generic.go:334] "Generic (PLEG): container finished" podID="c913d43d-9615-4436-ae90-56c68245d4a6" containerID="a90505b7ed2e3df50fe100275bb67208e4ee4f6c61dc7829fe25dec72dacfa4f" exitCode=0 Jan 06 15:20:55 crc kubenswrapper[4744]: I0106 15:20:55.930476 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" event={"ID":"c913d43d-9615-4436-ae90-56c68245d4a6","Type":"ContainerDied","Data":"a90505b7ed2e3df50fe100275bb67208e4ee4f6c61dc7829fe25dec72dacfa4f"} Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.490915 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.606404 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-nova-metadata-neutron-config-0\") pod \"c913d43d-9615-4436-ae90-56c68245d4a6\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.606615 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-metadata-combined-ca-bundle\") pod \"c913d43d-9615-4436-ae90-56c68245d4a6\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.607412 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-inventory\") pod \"c913d43d-9615-4436-ae90-56c68245d4a6\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.607570 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-ovn-metadata-agent-neutron-config-0\") pod \"c913d43d-9615-4436-ae90-56c68245d4a6\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.607708 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-ssh-key-openstack-edpm-ipam\") pod \"c913d43d-9615-4436-ae90-56c68245d4a6\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.607829 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psrt2\" (UniqueName: \"kubernetes.io/projected/c913d43d-9615-4436-ae90-56c68245d4a6-kube-api-access-psrt2\") pod \"c913d43d-9615-4436-ae90-56c68245d4a6\" (UID: \"c913d43d-9615-4436-ae90-56c68245d4a6\") " Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.612963 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c913d43d-9615-4436-ae90-56c68245d4a6-kube-api-access-psrt2" (OuterVolumeSpecName: "kube-api-access-psrt2") pod "c913d43d-9615-4436-ae90-56c68245d4a6" (UID: "c913d43d-9615-4436-ae90-56c68245d4a6"). InnerVolumeSpecName "kube-api-access-psrt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.627928 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c913d43d-9615-4436-ae90-56c68245d4a6" (UID: "c913d43d-9615-4436-ae90-56c68245d4a6"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.664108 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "c913d43d-9615-4436-ae90-56c68245d4a6" (UID: "c913d43d-9615-4436-ae90-56c68245d4a6"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.664293 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-inventory" (OuterVolumeSpecName: "inventory") pod "c913d43d-9615-4436-ae90-56c68245d4a6" (UID: "c913d43d-9615-4436-ae90-56c68245d4a6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.672552 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c913d43d-9615-4436-ae90-56c68245d4a6" (UID: "c913d43d-9615-4436-ae90-56c68245d4a6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.697321 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "c913d43d-9615-4436-ae90-56c68245d4a6" (UID: "c913d43d-9615-4436-ae90-56c68245d4a6"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.712302 4744 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.712639 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.712675 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.712691 4744 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.712962 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c913d43d-9615-4436-ae90-56c68245d4a6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.712984 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psrt2\" (UniqueName: \"kubernetes.io/projected/c913d43d-9615-4436-ae90-56c68245d4a6-kube-api-access-psrt2\") on node \"crc\" DevicePath \"\"" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.951611 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" event={"ID":"c913d43d-9615-4436-ae90-56c68245d4a6","Type":"ContainerDied","Data":"1763b73e61111243b3311e2d33d71752bcd2aa9aab815641f287f46148c49ef7"} Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.951950 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1763b73e61111243b3311e2d33d71752bcd2aa9aab815641f287f46148c49ef7" Jan 06 15:20:57 crc kubenswrapper[4744]: I0106 15:20:57.951727 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.058753 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4"] Jan 06 15:20:58 crc kubenswrapper[4744]: E0106 15:20:58.059426 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="registry-server" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059449 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="registry-server" Jan 06 15:20:58 crc kubenswrapper[4744]: E0106 15:20:58.059461 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerName="extract-content" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059470 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerName="extract-content" Jan 06 15:20:58 crc kubenswrapper[4744]: E0106 15:20:58.059515 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="extract-utilities" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059524 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="extract-utilities" Jan 06 15:20:58 crc kubenswrapper[4744]: E0106 15:20:58.059538 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerName="extract-utilities" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059545 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerName="extract-utilities" Jan 06 15:20:58 crc kubenswrapper[4744]: E0106 15:20:58.059561 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="extract-content" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059569 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="extract-content" Jan 06 15:20:58 crc kubenswrapper[4744]: E0106 15:20:58.059586 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c913d43d-9615-4436-ae90-56c68245d4a6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059595 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c913d43d-9615-4436-ae90-56c68245d4a6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 06 15:20:58 crc kubenswrapper[4744]: E0106 15:20:58.059615 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerName="registry-server" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059625 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerName="registry-server" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059886 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c913d43d-9615-4436-ae90-56c68245d4a6" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059908 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d37f7f03-5fa8-4b00-871c-4f43aac4026f" containerName="registry-server" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.059936 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ada8e8c-6c87-441e-8a39-cca5d58b3da0" containerName="registry-server" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.061239 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.065661 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.065670 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.065769 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.066265 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.066822 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.069041 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4"] Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.222357 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.222681 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26dw2\" (UniqueName: \"kubernetes.io/projected/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-kube-api-access-26dw2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.222782 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.222911 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.223021 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.325564 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.325624 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26dw2\" (UniqueName: \"kubernetes.io/projected/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-kube-api-access-26dw2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.325663 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.325757 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.325824 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.330503 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.330913 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.331673 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.331683 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.347757 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26dw2\" (UniqueName: \"kubernetes.io/projected/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-kube-api-access-26dw2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:58 crc kubenswrapper[4744]: I0106 15:20:58.383283 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:20:59 crc kubenswrapper[4744]: I0106 15:20:59.074846 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4"] Jan 06 15:20:59 crc kubenswrapper[4744]: W0106 15:20:59.075571 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d8bb7f8_6093_4f4a_81a1_6917f83346bf.slice/crio-da000dbb767e185b3266aa52e8cb95fac4c4f57e87458d9540bbd67668b15a84 WatchSource:0}: Error finding container da000dbb767e185b3266aa52e8cb95fac4c4f57e87458d9540bbd67668b15a84: Status 404 returned error can't find the container with id da000dbb767e185b3266aa52e8cb95fac4c4f57e87458d9540bbd67668b15a84 Jan 06 15:20:59 crc kubenswrapper[4744]: I0106 15:20:59.976079 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" event={"ID":"5d8bb7f8-6093-4f4a-81a1-6917f83346bf","Type":"ContainerStarted","Data":"da000dbb767e185b3266aa52e8cb95fac4c4f57e87458d9540bbd67668b15a84"} Jan 06 15:21:00 crc kubenswrapper[4744]: I0106 15:21:00.992146 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" event={"ID":"5d8bb7f8-6093-4f4a-81a1-6917f83346bf","Type":"ContainerStarted","Data":"25f7a5ba29a8d383b747bb0dbf9704c8dceaa154c530fb119818c036005299c3"} Jan 06 15:21:01 crc kubenswrapper[4744]: I0106 15:21:01.025402 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" podStartSLOduration=2.277806591 podStartE2EDuration="3.025381889s" podCreationTimestamp="2026-01-06 15:20:58 +0000 UTC" firstStartedPulling="2026-01-06 15:20:59.078320859 +0000 UTC m=+2655.705787177" lastFinishedPulling="2026-01-06 15:20:59.825896147 +0000 UTC m=+2656.453362475" observedRunningTime="2026-01-06 15:21:01.015041584 +0000 UTC m=+2657.642507902" watchObservedRunningTime="2026-01-06 15:21:01.025381889 +0000 UTC m=+2657.652848207" Jan 06 15:21:16 crc kubenswrapper[4744]: I0106 15:21:14.424531 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:21:16 crc kubenswrapper[4744]: I0106 15:21:14.425334 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.070575 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m29hj"] Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.073912 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.083875 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m29hj"] Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.191530 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-utilities\") pod \"community-operators-m29hj\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.191810 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7md6j\" (UniqueName: \"kubernetes.io/projected/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-kube-api-access-7md6j\") pod \"community-operators-m29hj\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.191863 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-catalog-content\") pod \"community-operators-m29hj\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.294339 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7md6j\" (UniqueName: \"kubernetes.io/projected/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-kube-api-access-7md6j\") pod \"community-operators-m29hj\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.294762 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-catalog-content\") pod \"community-operators-m29hj\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.295071 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-utilities\") pod \"community-operators-m29hj\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.295152 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-catalog-content\") pod \"community-operators-m29hj\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.295959 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-utilities\") pod \"community-operators-m29hj\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.313624 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7md6j\" (UniqueName: \"kubernetes.io/projected/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-kube-api-access-7md6j\") pod \"community-operators-m29hj\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.395973 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:36 crc kubenswrapper[4744]: W0106 15:21:36.938371 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad4b322b_43cd_4da9_92cd_73dd3be47ff6.slice/crio-a4c28060b4714df87f923f63cee4021f4a9cb448e4dbcfc41b885930d9d61a4f WatchSource:0}: Error finding container a4c28060b4714df87f923f63cee4021f4a9cb448e4dbcfc41b885930d9d61a4f: Status 404 returned error can't find the container with id a4c28060b4714df87f923f63cee4021f4a9cb448e4dbcfc41b885930d9d61a4f Jan 06 15:21:36 crc kubenswrapper[4744]: I0106 15:21:36.940198 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m29hj"] Jan 06 15:21:37 crc kubenswrapper[4744]: I0106 15:21:37.946639 4744 generic.go:334] "Generic (PLEG): container finished" podID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerID="5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b" exitCode=0 Jan 06 15:21:37 crc kubenswrapper[4744]: I0106 15:21:37.947326 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m29hj" event={"ID":"ad4b322b-43cd-4da9-92cd-73dd3be47ff6","Type":"ContainerDied","Data":"5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b"} Jan 06 15:21:37 crc kubenswrapper[4744]: I0106 15:21:37.947367 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m29hj" event={"ID":"ad4b322b-43cd-4da9-92cd-73dd3be47ff6","Type":"ContainerStarted","Data":"a4c28060b4714df87f923f63cee4021f4a9cb448e4dbcfc41b885930d9d61a4f"} Jan 06 15:21:39 crc kubenswrapper[4744]: I0106 15:21:39.972011 4744 generic.go:334] "Generic (PLEG): container finished" podID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerID="48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312" exitCode=0 Jan 06 15:21:39 crc kubenswrapper[4744]: I0106 15:21:39.972070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m29hj" event={"ID":"ad4b322b-43cd-4da9-92cd-73dd3be47ff6","Type":"ContainerDied","Data":"48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312"} Jan 06 15:21:41 crc kubenswrapper[4744]: I0106 15:21:41.001583 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m29hj" event={"ID":"ad4b322b-43cd-4da9-92cd-73dd3be47ff6","Type":"ContainerStarted","Data":"2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c"} Jan 06 15:21:41 crc kubenswrapper[4744]: I0106 15:21:41.047594 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m29hj" podStartSLOduration=2.506986619 podStartE2EDuration="5.04756339s" podCreationTimestamp="2026-01-06 15:21:36 +0000 UTC" firstStartedPulling="2026-01-06 15:21:37.951567621 +0000 UTC m=+2694.579033959" lastFinishedPulling="2026-01-06 15:21:40.492144412 +0000 UTC m=+2697.119610730" observedRunningTime="2026-01-06 15:21:41.026228373 +0000 UTC m=+2697.653694721" watchObservedRunningTime="2026-01-06 15:21:41.04756339 +0000 UTC m=+2697.675029728" Jan 06 15:21:44 crc kubenswrapper[4744]: I0106 15:21:44.423769 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:21:44 crc kubenswrapper[4744]: I0106 15:21:44.424458 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:21:44 crc kubenswrapper[4744]: I0106 15:21:44.424530 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:21:44 crc kubenswrapper[4744]: I0106 15:21:44.425988 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be2a215fc7aa9e274d639fb1b33f36c687f7a1d28458e250954de94d41c49b35"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:21:44 crc kubenswrapper[4744]: I0106 15:21:44.426127 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://be2a215fc7aa9e274d639fb1b33f36c687f7a1d28458e250954de94d41c49b35" gracePeriod=600 Jan 06 15:21:46 crc kubenswrapper[4744]: I0106 15:21:46.077333 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="be2a215fc7aa9e274d639fb1b33f36c687f7a1d28458e250954de94d41c49b35" exitCode=0 Jan 06 15:21:46 crc kubenswrapper[4744]: I0106 15:21:46.077439 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"be2a215fc7aa9e274d639fb1b33f36c687f7a1d28458e250954de94d41c49b35"} Jan 06 15:21:46 crc kubenswrapper[4744]: I0106 15:21:46.077966 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb"} Jan 06 15:21:46 crc kubenswrapper[4744]: I0106 15:21:46.077991 4744 scope.go:117] "RemoveContainer" containerID="2e87cbcee256c3f8755eab6682e7612e73f42d02e0329e95ca114add898bf8ce" Jan 06 15:21:46 crc kubenswrapper[4744]: I0106 15:21:46.397030 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:46 crc kubenswrapper[4744]: I0106 15:21:46.397315 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:46 crc kubenswrapper[4744]: I0106 15:21:46.451431 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:47 crc kubenswrapper[4744]: I0106 15:21:47.165018 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:47 crc kubenswrapper[4744]: I0106 15:21:47.250149 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m29hj"] Jan 06 15:21:49 crc kubenswrapper[4744]: I0106 15:21:49.127775 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-m29hj" podUID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerName="registry-server" containerID="cri-o://2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c" gracePeriod=2 Jan 06 15:21:49 crc kubenswrapper[4744]: I0106 15:21:49.742961 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:49 crc kubenswrapper[4744]: I0106 15:21:49.914914 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7md6j\" (UniqueName: \"kubernetes.io/projected/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-kube-api-access-7md6j\") pod \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " Jan 06 15:21:49 crc kubenswrapper[4744]: I0106 15:21:49.915224 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-utilities\") pod \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " Jan 06 15:21:49 crc kubenswrapper[4744]: I0106 15:21:49.915263 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-catalog-content\") pod \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\" (UID: \"ad4b322b-43cd-4da9-92cd-73dd3be47ff6\") " Jan 06 15:21:49 crc kubenswrapper[4744]: I0106 15:21:49.915953 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-utilities" (OuterVolumeSpecName: "utilities") pod "ad4b322b-43cd-4da9-92cd-73dd3be47ff6" (UID: "ad4b322b-43cd-4da9-92cd-73dd3be47ff6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:21:49 crc kubenswrapper[4744]: I0106 15:21:49.927376 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-kube-api-access-7md6j" (OuterVolumeSpecName: "kube-api-access-7md6j") pod "ad4b322b-43cd-4da9-92cd-73dd3be47ff6" (UID: "ad4b322b-43cd-4da9-92cd-73dd3be47ff6"). InnerVolumeSpecName "kube-api-access-7md6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.018145 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.018407 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7md6j\" (UniqueName: \"kubernetes.io/projected/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-kube-api-access-7md6j\") on node \"crc\" DevicePath \"\"" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.122363 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad4b322b-43cd-4da9-92cd-73dd3be47ff6" (UID: "ad4b322b-43cd-4da9-92cd-73dd3be47ff6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.145622 4744 generic.go:334] "Generic (PLEG): container finished" podID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerID="2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c" exitCode=0 Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.145667 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m29hj" event={"ID":"ad4b322b-43cd-4da9-92cd-73dd3be47ff6","Type":"ContainerDied","Data":"2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c"} Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.145706 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m29hj" event={"ID":"ad4b322b-43cd-4da9-92cd-73dd3be47ff6","Type":"ContainerDied","Data":"a4c28060b4714df87f923f63cee4021f4a9cb448e4dbcfc41b885930d9d61a4f"} Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.145733 4744 scope.go:117] "RemoveContainer" containerID="2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.145737 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m29hj" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.170704 4744 scope.go:117] "RemoveContainer" containerID="48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.188324 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-m29hj"] Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.203014 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-m29hj"] Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.206458 4744 scope.go:117] "RemoveContainer" containerID="5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.223303 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad4b322b-43cd-4da9-92cd-73dd3be47ff6-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.259797 4744 scope.go:117] "RemoveContainer" containerID="2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c" Jan 06 15:21:50 crc kubenswrapper[4744]: E0106 15:21:50.260236 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c\": container with ID starting with 2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c not found: ID does not exist" containerID="2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.260288 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c"} err="failed to get container status \"2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c\": rpc error: code = NotFound desc = could not find container \"2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c\": container with ID starting with 2303411c40f03b28adc69c9cd829bd498639e3fbfc04fd4b7fec18fe088f984c not found: ID does not exist" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.260315 4744 scope.go:117] "RemoveContainer" containerID="48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312" Jan 06 15:21:50 crc kubenswrapper[4744]: E0106 15:21:50.260665 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312\": container with ID starting with 48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312 not found: ID does not exist" containerID="48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.260719 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312"} err="failed to get container status \"48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312\": rpc error: code = NotFound desc = could not find container \"48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312\": container with ID starting with 48a8827c0c4b4267ffde50f57718c34449f53ced53bc8f3b2bfff1891af21312 not found: ID does not exist" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.260749 4744 scope.go:117] "RemoveContainer" containerID="5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b" Jan 06 15:21:50 crc kubenswrapper[4744]: E0106 15:21:50.261074 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b\": container with ID starting with 5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b not found: ID does not exist" containerID="5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b" Jan 06 15:21:50 crc kubenswrapper[4744]: I0106 15:21:50.261107 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b"} err="failed to get container status \"5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b\": rpc error: code = NotFound desc = could not find container \"5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b\": container with ID starting with 5ac67cebc0148530bf90847c6efe1e686b83af5542ebf210d9b7144b5d5b0c7b not found: ID does not exist" Jan 06 15:21:51 crc kubenswrapper[4744]: I0106 15:21:51.732045 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" path="/var/lib/kubelet/pods/ad4b322b-43cd-4da9-92cd-73dd3be47ff6/volumes" Jan 06 15:24:14 crc kubenswrapper[4744]: I0106 15:24:14.423679 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:24:14 crc kubenswrapper[4744]: I0106 15:24:14.424377 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:24:44 crc kubenswrapper[4744]: I0106 15:24:44.424391 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:24:44 crc kubenswrapper[4744]: I0106 15:24:44.426295 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:25:14 crc kubenswrapper[4744]: I0106 15:25:14.423655 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:25:14 crc kubenswrapper[4744]: I0106 15:25:14.424229 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:25:14 crc kubenswrapper[4744]: I0106 15:25:14.424306 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:25:14 crc kubenswrapper[4744]: I0106 15:25:14.425371 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:25:14 crc kubenswrapper[4744]: I0106 15:25:14.425430 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" gracePeriod=600 Jan 06 15:25:14 crc kubenswrapper[4744]: E0106 15:25:14.562186 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:25:14 crc kubenswrapper[4744]: I0106 15:25:14.770869 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" exitCode=0 Jan 06 15:25:14 crc kubenswrapper[4744]: I0106 15:25:14.770940 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb"} Jan 06 15:25:14 crc kubenswrapper[4744]: I0106 15:25:14.771278 4744 scope.go:117] "RemoveContainer" containerID="be2a215fc7aa9e274d639fb1b33f36c687f7a1d28458e250954de94d41c49b35" Jan 06 15:25:14 crc kubenswrapper[4744]: I0106 15:25:14.772215 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:25:14 crc kubenswrapper[4744]: E0106 15:25:14.772674 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:25:25 crc kubenswrapper[4744]: I0106 15:25:25.711097 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:25:25 crc kubenswrapper[4744]: E0106 15:25:25.711948 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:25:40 crc kubenswrapper[4744]: I0106 15:25:40.710968 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:25:40 crc kubenswrapper[4744]: E0106 15:25:40.711871 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:25:52 crc kubenswrapper[4744]: I0106 15:25:52.710949 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:25:52 crc kubenswrapper[4744]: E0106 15:25:52.711601 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:25:53 crc kubenswrapper[4744]: I0106 15:25:53.215673 4744 generic.go:334] "Generic (PLEG): container finished" podID="5d8bb7f8-6093-4f4a-81a1-6917f83346bf" containerID="25f7a5ba29a8d383b747bb0dbf9704c8dceaa154c530fb119818c036005299c3" exitCode=0 Jan 06 15:25:53 crc kubenswrapper[4744]: I0106 15:25:53.215761 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" event={"ID":"5d8bb7f8-6093-4f4a-81a1-6917f83346bf","Type":"ContainerDied","Data":"25f7a5ba29a8d383b747bb0dbf9704c8dceaa154c530fb119818c036005299c3"} Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.752041 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.889660 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-ssh-key-openstack-edpm-ipam\") pod \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.889763 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26dw2\" (UniqueName: \"kubernetes.io/projected/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-kube-api-access-26dw2\") pod \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.889867 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-secret-0\") pod \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.889999 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-combined-ca-bundle\") pod \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.890147 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-inventory\") pod \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\" (UID: \"5d8bb7f8-6093-4f4a-81a1-6917f83346bf\") " Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.897586 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-kube-api-access-26dw2" (OuterVolumeSpecName: "kube-api-access-26dw2") pod "5d8bb7f8-6093-4f4a-81a1-6917f83346bf" (UID: "5d8bb7f8-6093-4f4a-81a1-6917f83346bf"). InnerVolumeSpecName "kube-api-access-26dw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.898515 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5d8bb7f8-6093-4f4a-81a1-6917f83346bf" (UID: "5d8bb7f8-6093-4f4a-81a1-6917f83346bf"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.930694 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "5d8bb7f8-6093-4f4a-81a1-6917f83346bf" (UID: "5d8bb7f8-6093-4f4a-81a1-6917f83346bf"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.954474 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5d8bb7f8-6093-4f4a-81a1-6917f83346bf" (UID: "5d8bb7f8-6093-4f4a-81a1-6917f83346bf"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.956426 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-inventory" (OuterVolumeSpecName: "inventory") pod "5d8bb7f8-6093-4f4a-81a1-6917f83346bf" (UID: "5d8bb7f8-6093-4f4a-81a1-6917f83346bf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.993899 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26dw2\" (UniqueName: \"kubernetes.io/projected/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-kube-api-access-26dw2\") on node \"crc\" DevicePath \"\"" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.993941 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.993963 4744 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.993976 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:25:54 crc kubenswrapper[4744]: I0106 15:25:54.993987 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5d8bb7f8-6093-4f4a-81a1-6917f83346bf-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.262315 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" event={"ID":"5d8bb7f8-6093-4f4a-81a1-6917f83346bf","Type":"ContainerDied","Data":"da000dbb767e185b3266aa52e8cb95fac4c4f57e87458d9540bbd67668b15a84"} Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.262357 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da000dbb767e185b3266aa52e8cb95fac4c4f57e87458d9540bbd67668b15a84" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.262409 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.332747 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr"] Jan 06 15:25:55 crc kubenswrapper[4744]: E0106 15:25:55.333465 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerName="registry-server" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.333538 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerName="registry-server" Jan 06 15:25:55 crc kubenswrapper[4744]: E0106 15:25:55.333638 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerName="extract-utilities" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.333689 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerName="extract-utilities" Jan 06 15:25:55 crc kubenswrapper[4744]: E0106 15:25:55.333766 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d8bb7f8-6093-4f4a-81a1-6917f83346bf" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.333823 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d8bb7f8-6093-4f4a-81a1-6917f83346bf" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 06 15:25:55 crc kubenswrapper[4744]: E0106 15:25:55.334112 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerName="extract-content" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.334193 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerName="extract-content" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.334492 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d8bb7f8-6093-4f4a-81a1-6917f83346bf" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.334602 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad4b322b-43cd-4da9-92cd-73dd3be47ff6" containerName="registry-server" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.339043 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.342405 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.342636 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.342913 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.343067 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.343241 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.347077 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.347500 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.370016 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr"] Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.506055 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.506132 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.506685 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.506869 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.507022 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.507059 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmf8x\" (UniqueName: \"kubernetes.io/projected/af287382-0a8b-4a8c-a3a3-fbb41c6da499-kube-api-access-hmf8x\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.507113 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.507317 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.507412 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.609938 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.610373 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.610432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmf8x\" (UniqueName: \"kubernetes.io/projected/af287382-0a8b-4a8c-a3a3-fbb41c6da499-kube-api-access-hmf8x\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.610465 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.610499 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.610593 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.610689 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.610766 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.610828 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.612189 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.614923 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.615214 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.616813 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.618337 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.618389 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.621850 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.622283 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.633094 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmf8x\" (UniqueName: \"kubernetes.io/projected/af287382-0a8b-4a8c-a3a3-fbb41c6da499-kube-api-access-hmf8x\") pod \"nova-edpm-deployment-openstack-edpm-ipam-kdmcr\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:55 crc kubenswrapper[4744]: I0106 15:25:55.695761 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:25:56 crc kubenswrapper[4744]: I0106 15:25:56.244322 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr"] Jan 06 15:25:56 crc kubenswrapper[4744]: I0106 15:25:56.247760 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 15:25:56 crc kubenswrapper[4744]: I0106 15:25:56.271570 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" event={"ID":"af287382-0a8b-4a8c-a3a3-fbb41c6da499","Type":"ContainerStarted","Data":"43f6923481d91d293bd4feac9f177ff03ae8c6994492956386aa87e29c80f44b"} Jan 06 15:25:57 crc kubenswrapper[4744]: I0106 15:25:57.281441 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" event={"ID":"af287382-0a8b-4a8c-a3a3-fbb41c6da499","Type":"ContainerStarted","Data":"a2f699289a29815bc930bc9220548c708b9dd8aa9e3c3ec4d8e3f2c7a2187b11"} Jan 06 15:25:57 crc kubenswrapper[4744]: I0106 15:25:57.297210 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" podStartSLOduration=1.796673298 podStartE2EDuration="2.297191451s" podCreationTimestamp="2026-01-06 15:25:55 +0000 UTC" firstStartedPulling="2026-01-06 15:25:56.247450246 +0000 UTC m=+2952.874916564" lastFinishedPulling="2026-01-06 15:25:56.747968399 +0000 UTC m=+2953.375434717" observedRunningTime="2026-01-06 15:25:57.296351059 +0000 UTC m=+2953.923817387" watchObservedRunningTime="2026-01-06 15:25:57.297191451 +0000 UTC m=+2953.924657769" Jan 06 15:26:04 crc kubenswrapper[4744]: I0106 15:26:04.711792 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:26:04 crc kubenswrapper[4744]: E0106 15:26:04.712473 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:26:15 crc kubenswrapper[4744]: I0106 15:26:15.711287 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:26:15 crc kubenswrapper[4744]: E0106 15:26:15.712021 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:26:29 crc kubenswrapper[4744]: I0106 15:26:29.711905 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:26:29 crc kubenswrapper[4744]: E0106 15:26:29.713044 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:26:42 crc kubenswrapper[4744]: I0106 15:26:42.711528 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:26:42 crc kubenswrapper[4744]: E0106 15:26:42.712563 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:26:54 crc kubenswrapper[4744]: I0106 15:26:54.711582 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:26:54 crc kubenswrapper[4744]: E0106 15:26:54.712448 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:27:08 crc kubenswrapper[4744]: I0106 15:27:08.711688 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:27:08 crc kubenswrapper[4744]: E0106 15:27:08.712383 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:27:22 crc kubenswrapper[4744]: I0106 15:27:22.712807 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:27:22 crc kubenswrapper[4744]: E0106 15:27:22.713600 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:27:37 crc kubenswrapper[4744]: I0106 15:27:37.711379 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:27:37 crc kubenswrapper[4744]: E0106 15:27:37.713916 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:27:48 crc kubenswrapper[4744]: I0106 15:27:48.712278 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:27:48 crc kubenswrapper[4744]: E0106 15:27:48.713694 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:28:02 crc kubenswrapper[4744]: I0106 15:28:02.712818 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:28:02 crc kubenswrapper[4744]: E0106 15:28:02.713914 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:28:14 crc kubenswrapper[4744]: I0106 15:28:14.711820 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:28:14 crc kubenswrapper[4744]: E0106 15:28:14.713301 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:28:27 crc kubenswrapper[4744]: I0106 15:28:27.057920 4744 generic.go:334] "Generic (PLEG): container finished" podID="af287382-0a8b-4a8c-a3a3-fbb41c6da499" containerID="a2f699289a29815bc930bc9220548c708b9dd8aa9e3c3ec4d8e3f2c7a2187b11" exitCode=0 Jan 06 15:28:27 crc kubenswrapper[4744]: I0106 15:28:27.057995 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" event={"ID":"af287382-0a8b-4a8c-a3a3-fbb41c6da499","Type":"ContainerDied","Data":"a2f699289a29815bc930bc9220548c708b9dd8aa9e3c3ec4d8e3f2c7a2187b11"} Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.616134 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.764210 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-0\") pod \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.764266 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-0\") pod \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.764307 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-ssh-key-openstack-edpm-ipam\") pod \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.764352 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmf8x\" (UniqueName: \"kubernetes.io/projected/af287382-0a8b-4a8c-a3a3-fbb41c6da499-kube-api-access-hmf8x\") pod \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.764475 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-inventory\") pod \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.764563 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-extra-config-0\") pod \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.764629 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-1\") pod \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.764796 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-1\") pod \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.764865 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-combined-ca-bundle\") pod \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\" (UID: \"af287382-0a8b-4a8c-a3a3-fbb41c6da499\") " Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.771423 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "af287382-0a8b-4a8c-a3a3-fbb41c6da499" (UID: "af287382-0a8b-4a8c-a3a3-fbb41c6da499"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.785683 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af287382-0a8b-4a8c-a3a3-fbb41c6da499-kube-api-access-hmf8x" (OuterVolumeSpecName: "kube-api-access-hmf8x") pod "af287382-0a8b-4a8c-a3a3-fbb41c6da499" (UID: "af287382-0a8b-4a8c-a3a3-fbb41c6da499"). InnerVolumeSpecName "kube-api-access-hmf8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.799412 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "af287382-0a8b-4a8c-a3a3-fbb41c6da499" (UID: "af287382-0a8b-4a8c-a3a3-fbb41c6da499"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.799612 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "af287382-0a8b-4a8c-a3a3-fbb41c6da499" (UID: "af287382-0a8b-4a8c-a3a3-fbb41c6da499"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.808041 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "af287382-0a8b-4a8c-a3a3-fbb41c6da499" (UID: "af287382-0a8b-4a8c-a3a3-fbb41c6da499"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.817528 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-inventory" (OuterVolumeSpecName: "inventory") pod "af287382-0a8b-4a8c-a3a3-fbb41c6da499" (UID: "af287382-0a8b-4a8c-a3a3-fbb41c6da499"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.820105 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "af287382-0a8b-4a8c-a3a3-fbb41c6da499" (UID: "af287382-0a8b-4a8c-a3a3-fbb41c6da499"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.820327 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "af287382-0a8b-4a8c-a3a3-fbb41c6da499" (UID: "af287382-0a8b-4a8c-a3a3-fbb41c6da499"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.829497 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "af287382-0a8b-4a8c-a3a3-fbb41c6da499" (UID: "af287382-0a8b-4a8c-a3a3-fbb41c6da499"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.871207 4744 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.871243 4744 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.871253 4744 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.871265 4744 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.871278 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.871290 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmf8x\" (UniqueName: \"kubernetes.io/projected/af287382-0a8b-4a8c-a3a3-fbb41c6da499-kube-api-access-hmf8x\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.871302 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.871314 4744 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:28 crc kubenswrapper[4744]: I0106 15:28:28.871327 4744 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/af287382-0a8b-4a8c-a3a3-fbb41c6da499-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.082351 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" event={"ID":"af287382-0a8b-4a8c-a3a3-fbb41c6da499","Type":"ContainerDied","Data":"43f6923481d91d293bd4feac9f177ff03ae8c6994492956386aa87e29c80f44b"} Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.082718 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43f6923481d91d293bd4feac9f177ff03ae8c6994492956386aa87e29c80f44b" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.082796 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-kdmcr" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.181817 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d"] Jan 06 15:28:29 crc kubenswrapper[4744]: E0106 15:28:29.182581 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af287382-0a8b-4a8c-a3a3-fbb41c6da499" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.182633 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="af287382-0a8b-4a8c-a3a3-fbb41c6da499" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.183149 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="af287382-0a8b-4a8c-a3a3-fbb41c6da499" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.184690 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.190842 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.190909 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.190942 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.190850 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.191199 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.193033 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d"] Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.279638 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9k9n\" (UniqueName: \"kubernetes.io/projected/62caf07e-9021-49bb-9e27-79df992fc4c5-kube-api-access-p9k9n\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.279689 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.279853 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.279889 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.280029 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.280101 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.280134 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.382447 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.382552 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.382584 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.382671 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9k9n\" (UniqueName: \"kubernetes.io/projected/62caf07e-9021-49bb-9e27-79df992fc4c5-kube-api-access-p9k9n\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.382704 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.382809 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.382832 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.391740 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.392048 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.399963 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.400454 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.400808 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.401714 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.405068 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9k9n\" (UniqueName: \"kubernetes.io/projected/62caf07e-9021-49bb-9e27-79df992fc4c5-kube-api-access-p9k9n\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.561358 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:28:29 crc kubenswrapper[4744]: I0106 15:28:29.714408 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:28:29 crc kubenswrapper[4744]: E0106 15:28:29.715120 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:28:30 crc kubenswrapper[4744]: I0106 15:28:30.188919 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d"] Jan 06 15:28:30 crc kubenswrapper[4744]: W0106 15:28:30.193498 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62caf07e_9021_49bb_9e27_79df992fc4c5.slice/crio-f57277ac0d9dd636a596d8d8302194850e406843f752155c29a40de879c141e2 WatchSource:0}: Error finding container f57277ac0d9dd636a596d8d8302194850e406843f752155c29a40de879c141e2: Status 404 returned error can't find the container with id f57277ac0d9dd636a596d8d8302194850e406843f752155c29a40de879c141e2 Jan 06 15:28:31 crc kubenswrapper[4744]: I0106 15:28:31.107757 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" event={"ID":"62caf07e-9021-49bb-9e27-79df992fc4c5","Type":"ContainerStarted","Data":"3161b236ae0f21876ece731fce95f974e2db474080dbe396ebf2e194e30ac6f4"} Jan 06 15:28:31 crc kubenswrapper[4744]: I0106 15:28:31.108070 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" event={"ID":"62caf07e-9021-49bb-9e27-79df992fc4c5","Type":"ContainerStarted","Data":"f57277ac0d9dd636a596d8d8302194850e406843f752155c29a40de879c141e2"} Jan 06 15:28:31 crc kubenswrapper[4744]: I0106 15:28:31.131556 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" podStartSLOduration=1.6720526850000001 podStartE2EDuration="2.131537734s" podCreationTimestamp="2026-01-06 15:28:29 +0000 UTC" firstStartedPulling="2026-01-06 15:28:30.196017686 +0000 UTC m=+3106.823484004" lastFinishedPulling="2026-01-06 15:28:30.655502725 +0000 UTC m=+3107.282969053" observedRunningTime="2026-01-06 15:28:31.127484126 +0000 UTC m=+3107.754950444" watchObservedRunningTime="2026-01-06 15:28:31.131537734 +0000 UTC m=+3107.759004052" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.236690 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7m892"] Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.242012 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.251138 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7m892"] Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.325091 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqln2\" (UniqueName: \"kubernetes.io/projected/fd996a54-d4a3-407f-8808-d84797d9fb35-kube-api-access-fqln2\") pod \"redhat-marketplace-7m892\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.325551 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-utilities\") pod \"redhat-marketplace-7m892\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.325610 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-catalog-content\") pod \"redhat-marketplace-7m892\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.427973 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqln2\" (UniqueName: \"kubernetes.io/projected/fd996a54-d4a3-407f-8808-d84797d9fb35-kube-api-access-fqln2\") pod \"redhat-marketplace-7m892\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.428121 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-utilities\") pod \"redhat-marketplace-7m892\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.428190 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-catalog-content\") pod \"redhat-marketplace-7m892\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.428732 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-utilities\") pod \"redhat-marketplace-7m892\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.428847 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-catalog-content\") pod \"redhat-marketplace-7m892\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.447013 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqln2\" (UniqueName: \"kubernetes.io/projected/fd996a54-d4a3-407f-8808-d84797d9fb35-kube-api-access-fqln2\") pod \"redhat-marketplace-7m892\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:41 crc kubenswrapper[4744]: I0106 15:28:41.572880 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:42 crc kubenswrapper[4744]: I0106 15:28:42.124757 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7m892"] Jan 06 15:28:42 crc kubenswrapper[4744]: I0106 15:28:42.245294 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7m892" event={"ID":"fd996a54-d4a3-407f-8808-d84797d9fb35","Type":"ContainerStarted","Data":"97d6b728dd38e8ef86d5c841d0060260c8ed392a5341095b913fbd201f9cee0b"} Jan 06 15:28:43 crc kubenswrapper[4744]: I0106 15:28:43.256415 4744 generic.go:334] "Generic (PLEG): container finished" podID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerID="899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f" exitCode=0 Jan 06 15:28:43 crc kubenswrapper[4744]: I0106 15:28:43.256665 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7m892" event={"ID":"fd996a54-d4a3-407f-8808-d84797d9fb35","Type":"ContainerDied","Data":"899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f"} Jan 06 15:28:43 crc kubenswrapper[4744]: I0106 15:28:43.729252 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:28:43 crc kubenswrapper[4744]: E0106 15:28:43.729824 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:28:44 crc kubenswrapper[4744]: I0106 15:28:44.275662 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7m892" event={"ID":"fd996a54-d4a3-407f-8808-d84797d9fb35","Type":"ContainerStarted","Data":"ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de"} Jan 06 15:28:45 crc kubenswrapper[4744]: I0106 15:28:45.286256 4744 generic.go:334] "Generic (PLEG): container finished" podID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerID="ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de" exitCode=0 Jan 06 15:28:45 crc kubenswrapper[4744]: I0106 15:28:45.286311 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7m892" event={"ID":"fd996a54-d4a3-407f-8808-d84797d9fb35","Type":"ContainerDied","Data":"ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de"} Jan 06 15:28:46 crc kubenswrapper[4744]: I0106 15:28:46.305441 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7m892" event={"ID":"fd996a54-d4a3-407f-8808-d84797d9fb35","Type":"ContainerStarted","Data":"187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f"} Jan 06 15:28:46 crc kubenswrapper[4744]: I0106 15:28:46.339991 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7m892" podStartSLOduration=2.844939122 podStartE2EDuration="5.339952673s" podCreationTimestamp="2026-01-06 15:28:41 +0000 UTC" firstStartedPulling="2026-01-06 15:28:43.258664491 +0000 UTC m=+3119.886130809" lastFinishedPulling="2026-01-06 15:28:45.753678022 +0000 UTC m=+3122.381144360" observedRunningTime="2026-01-06 15:28:46.324966653 +0000 UTC m=+3122.952433011" watchObservedRunningTime="2026-01-06 15:28:46.339952673 +0000 UTC m=+3122.967418991" Jan 06 15:28:51 crc kubenswrapper[4744]: I0106 15:28:51.573944 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:51 crc kubenswrapper[4744]: I0106 15:28:51.575249 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:51 crc kubenswrapper[4744]: I0106 15:28:51.633885 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:52 crc kubenswrapper[4744]: I0106 15:28:52.445519 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:52 crc kubenswrapper[4744]: I0106 15:28:52.510926 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7m892"] Jan 06 15:28:54 crc kubenswrapper[4744]: I0106 15:28:54.411074 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7m892" podUID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerName="registry-server" containerID="cri-o://187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f" gracePeriod=2 Jan 06 15:28:54 crc kubenswrapper[4744]: I0106 15:28:54.914781 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.012074 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-catalog-content\") pod \"fd996a54-d4a3-407f-8808-d84797d9fb35\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.012211 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqln2\" (UniqueName: \"kubernetes.io/projected/fd996a54-d4a3-407f-8808-d84797d9fb35-kube-api-access-fqln2\") pod \"fd996a54-d4a3-407f-8808-d84797d9fb35\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.012270 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-utilities\") pod \"fd996a54-d4a3-407f-8808-d84797d9fb35\" (UID: \"fd996a54-d4a3-407f-8808-d84797d9fb35\") " Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.013267 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-utilities" (OuterVolumeSpecName: "utilities") pod "fd996a54-d4a3-407f-8808-d84797d9fb35" (UID: "fd996a54-d4a3-407f-8808-d84797d9fb35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.019914 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd996a54-d4a3-407f-8808-d84797d9fb35-kube-api-access-fqln2" (OuterVolumeSpecName: "kube-api-access-fqln2") pod "fd996a54-d4a3-407f-8808-d84797d9fb35" (UID: "fd996a54-d4a3-407f-8808-d84797d9fb35"). InnerVolumeSpecName "kube-api-access-fqln2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.036512 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd996a54-d4a3-407f-8808-d84797d9fb35" (UID: "fd996a54-d4a3-407f-8808-d84797d9fb35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.118691 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.118752 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd996a54-d4a3-407f-8808-d84797d9fb35-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.118772 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqln2\" (UniqueName: \"kubernetes.io/projected/fd996a54-d4a3-407f-8808-d84797d9fb35-kube-api-access-fqln2\") on node \"crc\" DevicePath \"\"" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.423573 4744 generic.go:334] "Generic (PLEG): container finished" podID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerID="187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f" exitCode=0 Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.423642 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7m892" event={"ID":"fd996a54-d4a3-407f-8808-d84797d9fb35","Type":"ContainerDied","Data":"187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f"} Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.423693 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7m892" event={"ID":"fd996a54-d4a3-407f-8808-d84797d9fb35","Type":"ContainerDied","Data":"97d6b728dd38e8ef86d5c841d0060260c8ed392a5341095b913fbd201f9cee0b"} Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.423722 4744 scope.go:117] "RemoveContainer" containerID="187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.423649 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7m892" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.453308 4744 scope.go:117] "RemoveContainer" containerID="ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.461313 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7m892"] Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.479230 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7m892"] Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.485571 4744 scope.go:117] "RemoveContainer" containerID="899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.552035 4744 scope.go:117] "RemoveContainer" containerID="187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f" Jan 06 15:28:55 crc kubenswrapper[4744]: E0106 15:28:55.552613 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f\": container with ID starting with 187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f not found: ID does not exist" containerID="187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.552667 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f"} err="failed to get container status \"187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f\": rpc error: code = NotFound desc = could not find container \"187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f\": container with ID starting with 187b8e0d41417cf7437c087b393dd32017e45e2d1e789b36d7c16cff273df12f not found: ID does not exist" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.552699 4744 scope.go:117] "RemoveContainer" containerID="ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de" Jan 06 15:28:55 crc kubenswrapper[4744]: E0106 15:28:55.553181 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de\": container with ID starting with ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de not found: ID does not exist" containerID="ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.553222 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de"} err="failed to get container status \"ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de\": rpc error: code = NotFound desc = could not find container \"ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de\": container with ID starting with ddf72e52ac79dd228859b3c06b20327c1fd733af219aa981d474af876824a7de not found: ID does not exist" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.553251 4744 scope.go:117] "RemoveContainer" containerID="899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f" Jan 06 15:28:55 crc kubenswrapper[4744]: E0106 15:28:55.553601 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f\": container with ID starting with 899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f not found: ID does not exist" containerID="899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.553624 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f"} err="failed to get container status \"899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f\": rpc error: code = NotFound desc = could not find container \"899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f\": container with ID starting with 899060ddb3f1a46b4e3b42a418dd9eb8365fd2be043bc28c035a1bce3da3409f not found: ID does not exist" Jan 06 15:28:55 crc kubenswrapper[4744]: I0106 15:28:55.734313 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd996a54-d4a3-407f-8808-d84797d9fb35" path="/var/lib/kubelet/pods/fd996a54-d4a3-407f-8808-d84797d9fb35/volumes" Jan 06 15:28:56 crc kubenswrapper[4744]: I0106 15:28:56.712099 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:28:56 crc kubenswrapper[4744]: E0106 15:28:56.712439 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:29:09 crc kubenswrapper[4744]: I0106 15:29:09.711709 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:29:09 crc kubenswrapper[4744]: E0106 15:29:09.712776 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:29:24 crc kubenswrapper[4744]: I0106 15:29:24.711359 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:29:24 crc kubenswrapper[4744]: E0106 15:29:24.712238 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:29:38 crc kubenswrapper[4744]: I0106 15:29:38.711676 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:29:38 crc kubenswrapper[4744]: E0106 15:29:38.712505 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:29:49 crc kubenswrapper[4744]: I0106 15:29:49.711048 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:29:49 crc kubenswrapper[4744]: E0106 15:29:49.714690 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.170822 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k"] Jan 06 15:30:00 crc kubenswrapper[4744]: E0106 15:30:00.171701 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerName="extract-utilities" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.171715 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerName="extract-utilities" Jan 06 15:30:00 crc kubenswrapper[4744]: E0106 15:30:00.171742 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerName="extract-content" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.171748 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerName="extract-content" Jan 06 15:30:00 crc kubenswrapper[4744]: E0106 15:30:00.171756 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerName="registry-server" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.171762 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerName="registry-server" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.172004 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd996a54-d4a3-407f-8808-d84797d9fb35" containerName="registry-server" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.172737 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.174761 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.175059 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.180921 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k"] Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.225096 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdwsd\" (UniqueName: \"kubernetes.io/projected/4c7b5506-b57e-4203-85b7-0451d584b652-kube-api-access-bdwsd\") pod \"collect-profiles-29461890-vxh2k\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.225177 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c7b5506-b57e-4203-85b7-0451d584b652-config-volume\") pod \"collect-profiles-29461890-vxh2k\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.225256 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c7b5506-b57e-4203-85b7-0451d584b652-secret-volume\") pod \"collect-profiles-29461890-vxh2k\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.327367 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdwsd\" (UniqueName: \"kubernetes.io/projected/4c7b5506-b57e-4203-85b7-0451d584b652-kube-api-access-bdwsd\") pod \"collect-profiles-29461890-vxh2k\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.327433 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c7b5506-b57e-4203-85b7-0451d584b652-config-volume\") pod \"collect-profiles-29461890-vxh2k\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.327507 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c7b5506-b57e-4203-85b7-0451d584b652-secret-volume\") pod \"collect-profiles-29461890-vxh2k\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.328332 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c7b5506-b57e-4203-85b7-0451d584b652-config-volume\") pod \"collect-profiles-29461890-vxh2k\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.338077 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c7b5506-b57e-4203-85b7-0451d584b652-secret-volume\") pod \"collect-profiles-29461890-vxh2k\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.353608 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdwsd\" (UniqueName: \"kubernetes.io/projected/4c7b5506-b57e-4203-85b7-0451d584b652-kube-api-access-bdwsd\") pod \"collect-profiles-29461890-vxh2k\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:00 crc kubenswrapper[4744]: I0106 15:30:00.508959 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:01 crc kubenswrapper[4744]: I0106 15:30:01.019540 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k"] Jan 06 15:30:01 crc kubenswrapper[4744]: I0106 15:30:01.228645 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" event={"ID":"4c7b5506-b57e-4203-85b7-0451d584b652","Type":"ContainerStarted","Data":"2dede9b56059ba8afe86ff7853489d5dff6ffc9f1f0816a5e88b3e156436a3c4"} Jan 06 15:30:01 crc kubenswrapper[4744]: I0106 15:30:01.228686 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" event={"ID":"4c7b5506-b57e-4203-85b7-0451d584b652","Type":"ContainerStarted","Data":"def32402bc97b30581423c08d96dcb33c6c0e8692dac0e8dff9f12e26bf3681b"} Jan 06 15:30:01 crc kubenswrapper[4744]: I0106 15:30:01.255732 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" podStartSLOduration=1.255714945 podStartE2EDuration="1.255714945s" podCreationTimestamp="2026-01-06 15:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 15:30:01.243798728 +0000 UTC m=+3197.871265086" watchObservedRunningTime="2026-01-06 15:30:01.255714945 +0000 UTC m=+3197.883181263" Jan 06 15:30:02 crc kubenswrapper[4744]: I0106 15:30:02.245574 4744 generic.go:334] "Generic (PLEG): container finished" podID="4c7b5506-b57e-4203-85b7-0451d584b652" containerID="2dede9b56059ba8afe86ff7853489d5dff6ffc9f1f0816a5e88b3e156436a3c4" exitCode=0 Jan 06 15:30:02 crc kubenswrapper[4744]: I0106 15:30:02.245689 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" event={"ID":"4c7b5506-b57e-4203-85b7-0451d584b652","Type":"ContainerDied","Data":"2dede9b56059ba8afe86ff7853489d5dff6ffc9f1f0816a5e88b3e156436a3c4"} Jan 06 15:30:02 crc kubenswrapper[4744]: I0106 15:30:02.712627 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:30:02 crc kubenswrapper[4744]: E0106 15:30:02.713123 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:30:03 crc kubenswrapper[4744]: I0106 15:30:03.784896 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:03 crc kubenswrapper[4744]: I0106 15:30:03.929211 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c7b5506-b57e-4203-85b7-0451d584b652-config-volume\") pod \"4c7b5506-b57e-4203-85b7-0451d584b652\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " Jan 06 15:30:03 crc kubenswrapper[4744]: I0106 15:30:03.929474 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c7b5506-b57e-4203-85b7-0451d584b652-secret-volume\") pod \"4c7b5506-b57e-4203-85b7-0451d584b652\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " Jan 06 15:30:03 crc kubenswrapper[4744]: I0106 15:30:03.929554 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdwsd\" (UniqueName: \"kubernetes.io/projected/4c7b5506-b57e-4203-85b7-0451d584b652-kube-api-access-bdwsd\") pod \"4c7b5506-b57e-4203-85b7-0451d584b652\" (UID: \"4c7b5506-b57e-4203-85b7-0451d584b652\") " Jan 06 15:30:03 crc kubenswrapper[4744]: I0106 15:30:03.930171 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c7b5506-b57e-4203-85b7-0451d584b652-config-volume" (OuterVolumeSpecName: "config-volume") pod "4c7b5506-b57e-4203-85b7-0451d584b652" (UID: "4c7b5506-b57e-4203-85b7-0451d584b652"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:30:03 crc kubenswrapper[4744]: I0106 15:30:03.930373 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4c7b5506-b57e-4203-85b7-0451d584b652-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 15:30:03 crc kubenswrapper[4744]: I0106 15:30:03.935925 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c7b5506-b57e-4203-85b7-0451d584b652-kube-api-access-bdwsd" (OuterVolumeSpecName: "kube-api-access-bdwsd") pod "4c7b5506-b57e-4203-85b7-0451d584b652" (UID: "4c7b5506-b57e-4203-85b7-0451d584b652"). InnerVolumeSpecName "kube-api-access-bdwsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:30:03 crc kubenswrapper[4744]: I0106 15:30:03.946003 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7b5506-b57e-4203-85b7-0451d584b652-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4c7b5506-b57e-4203-85b7-0451d584b652" (UID: "4c7b5506-b57e-4203-85b7-0451d584b652"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:30:04 crc kubenswrapper[4744]: I0106 15:30:04.032979 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4c7b5506-b57e-4203-85b7-0451d584b652-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 06 15:30:04 crc kubenswrapper[4744]: I0106 15:30:04.033224 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdwsd\" (UniqueName: \"kubernetes.io/projected/4c7b5506-b57e-4203-85b7-0451d584b652-kube-api-access-bdwsd\") on node \"crc\" DevicePath \"\"" Jan 06 15:30:04 crc kubenswrapper[4744]: I0106 15:30:04.272472 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" event={"ID":"4c7b5506-b57e-4203-85b7-0451d584b652","Type":"ContainerDied","Data":"def32402bc97b30581423c08d96dcb33c6c0e8692dac0e8dff9f12e26bf3681b"} Jan 06 15:30:04 crc kubenswrapper[4744]: I0106 15:30:04.272524 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="def32402bc97b30581423c08d96dcb33c6c0e8692dac0e8dff9f12e26bf3681b" Jan 06 15:30:04 crc kubenswrapper[4744]: I0106 15:30:04.272588 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k" Jan 06 15:30:04 crc kubenswrapper[4744]: I0106 15:30:04.340368 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z"] Jan 06 15:30:04 crc kubenswrapper[4744]: I0106 15:30:04.368018 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461845-rd62z"] Jan 06 15:30:05 crc kubenswrapper[4744]: I0106 15:30:05.730667 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="277904fe-5db6-4bfa-82fe-dca1c673e5dc" path="/var/lib/kubelet/pods/277904fe-5db6-4bfa-82fe-dca1c673e5dc/volumes" Jan 06 15:30:12 crc kubenswrapper[4744]: I0106 15:30:12.434982 4744 scope.go:117] "RemoveContainer" containerID="d50d792cf16292e03b68b9b70c90fb64c21c89dc2b165dfe33694797822614e0" Jan 06 15:30:13 crc kubenswrapper[4744]: I0106 15:30:13.727736 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:30:13 crc kubenswrapper[4744]: E0106 15:30:13.729033 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:30:26 crc kubenswrapper[4744]: I0106 15:30:26.711829 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:30:27 crc kubenswrapper[4744]: I0106 15:30:27.547376 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"38d5508f3beb00f9bb486f7d949acf4397056b3e1d199506d1c8635be557972c"} Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.348812 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gst6w"] Jan 06 15:30:31 crc kubenswrapper[4744]: E0106 15:30:31.350174 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7b5506-b57e-4203-85b7-0451d584b652" containerName="collect-profiles" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.350193 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7b5506-b57e-4203-85b7-0451d584b652" containerName="collect-profiles" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.350501 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c7b5506-b57e-4203-85b7-0451d584b652" containerName="collect-profiles" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.352816 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.364267 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gst6w"] Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.532814 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgsch\" (UniqueName: \"kubernetes.io/projected/f4189077-1456-4a6a-992e-fd1cd6d56fcf-kube-api-access-fgsch\") pod \"certified-operators-gst6w\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.533357 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-utilities\") pod \"certified-operators-gst6w\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.533389 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-catalog-content\") pod \"certified-operators-gst6w\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.635964 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-utilities\") pod \"certified-operators-gst6w\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.636013 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-catalog-content\") pod \"certified-operators-gst6w\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.636263 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgsch\" (UniqueName: \"kubernetes.io/projected/f4189077-1456-4a6a-992e-fd1cd6d56fcf-kube-api-access-fgsch\") pod \"certified-operators-gst6w\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.636684 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-utilities\") pod \"certified-operators-gst6w\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.636883 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-catalog-content\") pod \"certified-operators-gst6w\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.659844 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgsch\" (UniqueName: \"kubernetes.io/projected/f4189077-1456-4a6a-992e-fd1cd6d56fcf-kube-api-access-fgsch\") pod \"certified-operators-gst6w\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:31 crc kubenswrapper[4744]: I0106 15:30:31.680183 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:32 crc kubenswrapper[4744]: I0106 15:30:32.358800 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gst6w"] Jan 06 15:30:32 crc kubenswrapper[4744]: I0106 15:30:32.609695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gst6w" event={"ID":"f4189077-1456-4a6a-992e-fd1cd6d56fcf","Type":"ContainerStarted","Data":"dc2c7be0acca2cebaa0a5307d94a3eb187abb30c653e1d0c656c8aa2b868aa00"} Jan 06 15:30:33 crc kubenswrapper[4744]: I0106 15:30:33.625067 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerID="85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1" exitCode=0 Jan 06 15:30:33 crc kubenswrapper[4744]: I0106 15:30:33.625210 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gst6w" event={"ID":"f4189077-1456-4a6a-992e-fd1cd6d56fcf","Type":"ContainerDied","Data":"85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1"} Jan 06 15:30:35 crc kubenswrapper[4744]: I0106 15:30:35.662625 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gst6w" event={"ID":"f4189077-1456-4a6a-992e-fd1cd6d56fcf","Type":"ContainerStarted","Data":"d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92"} Jan 06 15:30:36 crc kubenswrapper[4744]: I0106 15:30:36.675451 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerID="d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92" exitCode=0 Jan 06 15:30:36 crc kubenswrapper[4744]: I0106 15:30:36.675650 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gst6w" event={"ID":"f4189077-1456-4a6a-992e-fd1cd6d56fcf","Type":"ContainerDied","Data":"d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92"} Jan 06 15:30:37 crc kubenswrapper[4744]: I0106 15:30:37.688224 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gst6w" event={"ID":"f4189077-1456-4a6a-992e-fd1cd6d56fcf","Type":"ContainerStarted","Data":"ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a"} Jan 06 15:30:37 crc kubenswrapper[4744]: I0106 15:30:37.719667 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gst6w" podStartSLOduration=3.206480448 podStartE2EDuration="6.719633351s" podCreationTimestamp="2026-01-06 15:30:31 +0000 UTC" firstStartedPulling="2026-01-06 15:30:33.627941393 +0000 UTC m=+3230.255407751" lastFinishedPulling="2026-01-06 15:30:37.141094346 +0000 UTC m=+3233.768560654" observedRunningTime="2026-01-06 15:30:37.710121147 +0000 UTC m=+3234.337587475" watchObservedRunningTime="2026-01-06 15:30:37.719633351 +0000 UTC m=+3234.347099709" Jan 06 15:30:41 crc kubenswrapper[4744]: I0106 15:30:41.680656 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:41 crc kubenswrapper[4744]: I0106 15:30:41.681248 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:41 crc kubenswrapper[4744]: I0106 15:30:41.754072 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:51 crc kubenswrapper[4744]: I0106 15:30:51.775836 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:51 crc kubenswrapper[4744]: I0106 15:30:51.827822 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gst6w"] Jan 06 15:30:51 crc kubenswrapper[4744]: I0106 15:30:51.843639 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gst6w" podUID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerName="registry-server" containerID="cri-o://ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a" gracePeriod=2 Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.411609 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.492303 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-catalog-content\") pod \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.492556 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-utilities\") pod \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.492823 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgsch\" (UniqueName: \"kubernetes.io/projected/f4189077-1456-4a6a-992e-fd1cd6d56fcf-kube-api-access-fgsch\") pod \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\" (UID: \"f4189077-1456-4a6a-992e-fd1cd6d56fcf\") " Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.493221 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-utilities" (OuterVolumeSpecName: "utilities") pod "f4189077-1456-4a6a-992e-fd1cd6d56fcf" (UID: "f4189077-1456-4a6a-992e-fd1cd6d56fcf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.493928 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.502316 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4189077-1456-4a6a-992e-fd1cd6d56fcf-kube-api-access-fgsch" (OuterVolumeSpecName: "kube-api-access-fgsch") pod "f4189077-1456-4a6a-992e-fd1cd6d56fcf" (UID: "f4189077-1456-4a6a-992e-fd1cd6d56fcf"). InnerVolumeSpecName "kube-api-access-fgsch". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.543770 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4189077-1456-4a6a-992e-fd1cd6d56fcf" (UID: "f4189077-1456-4a6a-992e-fd1cd6d56fcf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.597112 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4189077-1456-4a6a-992e-fd1cd6d56fcf-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.597152 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgsch\" (UniqueName: \"kubernetes.io/projected/f4189077-1456-4a6a-992e-fd1cd6d56fcf-kube-api-access-fgsch\") on node \"crc\" DevicePath \"\"" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.869753 4744 generic.go:334] "Generic (PLEG): container finished" podID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerID="ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a" exitCode=0 Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.869797 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gst6w" event={"ID":"f4189077-1456-4a6a-992e-fd1cd6d56fcf","Type":"ContainerDied","Data":"ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a"} Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.869825 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gst6w" event={"ID":"f4189077-1456-4a6a-992e-fd1cd6d56fcf","Type":"ContainerDied","Data":"dc2c7be0acca2cebaa0a5307d94a3eb187abb30c653e1d0c656c8aa2b868aa00"} Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.869843 4744 scope.go:117] "RemoveContainer" containerID="ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.869871 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gst6w" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.921352 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gst6w"] Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.932558 4744 scope.go:117] "RemoveContainer" containerID="d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92" Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.936883 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gst6w"] Jan 06 15:30:52 crc kubenswrapper[4744]: I0106 15:30:52.961774 4744 scope.go:117] "RemoveContainer" containerID="85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1" Jan 06 15:30:53 crc kubenswrapper[4744]: I0106 15:30:53.020186 4744 scope.go:117] "RemoveContainer" containerID="ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a" Jan 06 15:30:53 crc kubenswrapper[4744]: E0106 15:30:53.020614 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a\": container with ID starting with ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a not found: ID does not exist" containerID="ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a" Jan 06 15:30:53 crc kubenswrapper[4744]: I0106 15:30:53.020659 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a"} err="failed to get container status \"ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a\": rpc error: code = NotFound desc = could not find container \"ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a\": container with ID starting with ae1ba229b57af0fb0f63de3eca00689298ad82e3f6bb52f70ff9cda67792588a not found: ID does not exist" Jan 06 15:30:53 crc kubenswrapper[4744]: I0106 15:30:53.020680 4744 scope.go:117] "RemoveContainer" containerID="d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92" Jan 06 15:30:53 crc kubenswrapper[4744]: E0106 15:30:53.021004 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92\": container with ID starting with d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92 not found: ID does not exist" containerID="d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92" Jan 06 15:30:53 crc kubenswrapper[4744]: I0106 15:30:53.021042 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92"} err="failed to get container status \"d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92\": rpc error: code = NotFound desc = could not find container \"d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92\": container with ID starting with d67dc829804eb0a90831480b5358d585257f77736c5cb3921535535997b77c92 not found: ID does not exist" Jan 06 15:30:53 crc kubenswrapper[4744]: I0106 15:30:53.021057 4744 scope.go:117] "RemoveContainer" containerID="85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1" Jan 06 15:30:53 crc kubenswrapper[4744]: E0106 15:30:53.021341 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1\": container with ID starting with 85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1 not found: ID does not exist" containerID="85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1" Jan 06 15:30:53 crc kubenswrapper[4744]: I0106 15:30:53.021360 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1"} err="failed to get container status \"85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1\": rpc error: code = NotFound desc = could not find container \"85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1\": container with ID starting with 85c8dd92a81bae8622f5b5d72254103566c66afdcbdd8f8affee755625c092b1 not found: ID does not exist" Jan 06 15:30:53 crc kubenswrapper[4744]: I0106 15:30:53.726073 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" path="/var/lib/kubelet/pods/f4189077-1456-4a6a-992e-fd1cd6d56fcf/volumes" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.460693 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lc5t7"] Jan 06 15:31:04 crc kubenswrapper[4744]: E0106 15:31:04.461655 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerName="registry-server" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.461670 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerName="registry-server" Jan 06 15:31:04 crc kubenswrapper[4744]: E0106 15:31:04.461697 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerName="extract-utilities" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.461705 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerName="extract-utilities" Jan 06 15:31:04 crc kubenswrapper[4744]: E0106 15:31:04.461724 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerName="extract-content" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.461732 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerName="extract-content" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.462051 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4189077-1456-4a6a-992e-fd1cd6d56fcf" containerName="registry-server" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.464912 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.478383 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lc5t7"] Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.510080 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgbw7\" (UniqueName: \"kubernetes.io/projected/91efe524-a251-4191-84f1-a611aec44891-kube-api-access-fgbw7\") pod \"redhat-operators-lc5t7\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.510373 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-catalog-content\") pod \"redhat-operators-lc5t7\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.510441 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-utilities\") pod \"redhat-operators-lc5t7\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.613288 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgbw7\" (UniqueName: \"kubernetes.io/projected/91efe524-a251-4191-84f1-a611aec44891-kube-api-access-fgbw7\") pod \"redhat-operators-lc5t7\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.613681 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-catalog-content\") pod \"redhat-operators-lc5t7\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.613722 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-utilities\") pod \"redhat-operators-lc5t7\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.614200 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-catalog-content\") pod \"redhat-operators-lc5t7\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.614236 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-utilities\") pod \"redhat-operators-lc5t7\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.646783 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgbw7\" (UniqueName: \"kubernetes.io/projected/91efe524-a251-4191-84f1-a611aec44891-kube-api-access-fgbw7\") pod \"redhat-operators-lc5t7\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:04 crc kubenswrapper[4744]: I0106 15:31:04.815829 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:05 crc kubenswrapper[4744]: I0106 15:31:05.377502 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lc5t7"] Jan 06 15:31:06 crc kubenswrapper[4744]: I0106 15:31:06.037423 4744 generic.go:334] "Generic (PLEG): container finished" podID="91efe524-a251-4191-84f1-a611aec44891" containerID="f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4" exitCode=0 Jan 06 15:31:06 crc kubenswrapper[4744]: I0106 15:31:06.037719 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc5t7" event={"ID":"91efe524-a251-4191-84f1-a611aec44891","Type":"ContainerDied","Data":"f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4"} Jan 06 15:31:06 crc kubenswrapper[4744]: I0106 15:31:06.037747 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc5t7" event={"ID":"91efe524-a251-4191-84f1-a611aec44891","Type":"ContainerStarted","Data":"1930b1c893ba066e1846d13b2f399f95718febbdacbbe84de73689de18ac0539"} Jan 06 15:31:06 crc kubenswrapper[4744]: I0106 15:31:06.040093 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 15:31:07 crc kubenswrapper[4744]: I0106 15:31:07.049668 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc5t7" event={"ID":"91efe524-a251-4191-84f1-a611aec44891","Type":"ContainerStarted","Data":"312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251"} Jan 06 15:31:12 crc kubenswrapper[4744]: I0106 15:31:12.100562 4744 generic.go:334] "Generic (PLEG): container finished" podID="91efe524-a251-4191-84f1-a611aec44891" containerID="312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251" exitCode=0 Jan 06 15:31:12 crc kubenswrapper[4744]: I0106 15:31:12.100635 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc5t7" event={"ID":"91efe524-a251-4191-84f1-a611aec44891","Type":"ContainerDied","Data":"312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251"} Jan 06 15:31:13 crc kubenswrapper[4744]: I0106 15:31:13.113780 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc5t7" event={"ID":"91efe524-a251-4191-84f1-a611aec44891","Type":"ContainerStarted","Data":"95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2"} Jan 06 15:31:13 crc kubenswrapper[4744]: I0106 15:31:13.145880 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lc5t7" podStartSLOduration=2.600060313 podStartE2EDuration="9.145861822s" podCreationTimestamp="2026-01-06 15:31:04 +0000 UTC" firstStartedPulling="2026-01-06 15:31:06.039887229 +0000 UTC m=+3262.667353537" lastFinishedPulling="2026-01-06 15:31:12.585688708 +0000 UTC m=+3269.213155046" observedRunningTime="2026-01-06 15:31:13.136922474 +0000 UTC m=+3269.764388792" watchObservedRunningTime="2026-01-06 15:31:13.145861822 +0000 UTC m=+3269.773328140" Jan 06 15:31:14 crc kubenswrapper[4744]: I0106 15:31:14.816196 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:14 crc kubenswrapper[4744]: I0106 15:31:14.816482 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:15 crc kubenswrapper[4744]: I0106 15:31:15.866058 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc5t7" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="registry-server" probeResult="failure" output=< Jan 06 15:31:15 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 15:31:15 crc kubenswrapper[4744]: > Jan 06 15:31:22 crc kubenswrapper[4744]: I0106 15:31:22.216604 4744 generic.go:334] "Generic (PLEG): container finished" podID="62caf07e-9021-49bb-9e27-79df992fc4c5" containerID="3161b236ae0f21876ece731fce95f974e2db474080dbe396ebf2e194e30ac6f4" exitCode=0 Jan 06 15:31:22 crc kubenswrapper[4744]: I0106 15:31:22.216713 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" event={"ID":"62caf07e-9021-49bb-9e27-79df992fc4c5","Type":"ContainerDied","Data":"3161b236ae0f21876ece731fce95f974e2db474080dbe396ebf2e194e30ac6f4"} Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.813688 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.949314 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-2\") pod \"62caf07e-9021-49bb-9e27-79df992fc4c5\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.949847 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-telemetry-combined-ca-bundle\") pod \"62caf07e-9021-49bb-9e27-79df992fc4c5\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.949975 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-0\") pod \"62caf07e-9021-49bb-9e27-79df992fc4c5\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.950412 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-inventory\") pod \"62caf07e-9021-49bb-9e27-79df992fc4c5\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.950535 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-1\") pod \"62caf07e-9021-49bb-9e27-79df992fc4c5\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.950720 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9k9n\" (UniqueName: \"kubernetes.io/projected/62caf07e-9021-49bb-9e27-79df992fc4c5-kube-api-access-p9k9n\") pod \"62caf07e-9021-49bb-9e27-79df992fc4c5\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.951100 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ssh-key-openstack-edpm-ipam\") pod \"62caf07e-9021-49bb-9e27-79df992fc4c5\" (UID: \"62caf07e-9021-49bb-9e27-79df992fc4c5\") " Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.955416 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "62caf07e-9021-49bb-9e27-79df992fc4c5" (UID: "62caf07e-9021-49bb-9e27-79df992fc4c5"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.960193 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62caf07e-9021-49bb-9e27-79df992fc4c5-kube-api-access-p9k9n" (OuterVolumeSpecName: "kube-api-access-p9k9n") pod "62caf07e-9021-49bb-9e27-79df992fc4c5" (UID: "62caf07e-9021-49bb-9e27-79df992fc4c5"). InnerVolumeSpecName "kube-api-access-p9k9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.983194 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "62caf07e-9021-49bb-9e27-79df992fc4c5" (UID: "62caf07e-9021-49bb-9e27-79df992fc4c5"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.984680 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-inventory" (OuterVolumeSpecName: "inventory") pod "62caf07e-9021-49bb-9e27-79df992fc4c5" (UID: "62caf07e-9021-49bb-9e27-79df992fc4c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:31:23 crc kubenswrapper[4744]: I0106 15:31:23.992348 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "62caf07e-9021-49bb-9e27-79df992fc4c5" (UID: "62caf07e-9021-49bb-9e27-79df992fc4c5"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.001797 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "62caf07e-9021-49bb-9e27-79df992fc4c5" (UID: "62caf07e-9021-49bb-9e27-79df992fc4c5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.014915 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "62caf07e-9021-49bb-9e27-79df992fc4c5" (UID: "62caf07e-9021-49bb-9e27-79df992fc4c5"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.053960 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.053998 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.054010 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9k9n\" (UniqueName: \"kubernetes.io/projected/62caf07e-9021-49bb-9e27-79df992fc4c5-kube-api-access-p9k9n\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.054022 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.054032 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.054043 4744 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.054053 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/62caf07e-9021-49bb-9e27-79df992fc4c5-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.241672 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" event={"ID":"62caf07e-9021-49bb-9e27-79df992fc4c5","Type":"ContainerDied","Data":"f57277ac0d9dd636a596d8d8302194850e406843f752155c29a40de879c141e2"} Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.241714 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f57277ac0d9dd636a596d8d8302194850e406843f752155c29a40de879c141e2" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.242236 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.373489 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr"] Jan 06 15:31:24 crc kubenswrapper[4744]: E0106 15:31:24.374356 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62caf07e-9021-49bb-9e27-79df992fc4c5" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.374372 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="62caf07e-9021-49bb-9e27-79df992fc4c5" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.374605 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="62caf07e-9021-49bb-9e27-79df992fc4c5" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.376812 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.380083 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.380124 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.380085 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.383655 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.383762 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.401357 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr"] Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.564045 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.564113 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.564204 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.564291 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dcvs\" (UniqueName: \"kubernetes.io/projected/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-kube-api-access-9dcvs\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.564396 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.564438 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ssh-key-openstack-edpm-ipam\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.564479 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.667635 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.667977 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.668055 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dcvs\" (UniqueName: \"kubernetes.io/projected/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-kube-api-access-9dcvs\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.668124 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.668190 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ssh-key-openstack-edpm-ipam\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.668258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.668490 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.672362 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.672820 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.672849 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.674650 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ssh-key-openstack-edpm-ipam\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.675239 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.678012 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.691354 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dcvs\" (UniqueName: \"kubernetes.io/projected/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-kube-api-access-9dcvs\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:24 crc kubenswrapper[4744]: I0106 15:31:24.706519 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:31:25 crc kubenswrapper[4744]: W0106 15:31:25.344077 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5216d2ec_d2e3_4f91_97d5_5839eaeb173d.slice/crio-b0fd8d7d7dbc8758425a04ebce9bbb76bd2e6d59f43c13c54c109ec308ccfb3b WatchSource:0}: Error finding container b0fd8d7d7dbc8758425a04ebce9bbb76bd2e6d59f43c13c54c109ec308ccfb3b: Status 404 returned error can't find the container with id b0fd8d7d7dbc8758425a04ebce9bbb76bd2e6d59f43c13c54c109ec308ccfb3b Jan 06 15:31:25 crc kubenswrapper[4744]: I0106 15:31:25.348149 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr"] Jan 06 15:31:25 crc kubenswrapper[4744]: I0106 15:31:25.901493 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lc5t7" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="registry-server" probeResult="failure" output=< Jan 06 15:31:25 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 15:31:25 crc kubenswrapper[4744]: > Jan 06 15:31:26 crc kubenswrapper[4744]: I0106 15:31:26.279909 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" event={"ID":"5216d2ec-d2e3-4f91-97d5-5839eaeb173d","Type":"ContainerStarted","Data":"0533b8addfaf4f5d766d5949829f6c13618dfe1a470614093b60085c9b5d1a38"} Jan 06 15:31:26 crc kubenswrapper[4744]: I0106 15:31:26.280204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" event={"ID":"5216d2ec-d2e3-4f91-97d5-5839eaeb173d","Type":"ContainerStarted","Data":"b0fd8d7d7dbc8758425a04ebce9bbb76bd2e6d59f43c13c54c109ec308ccfb3b"} Jan 06 15:31:26 crc kubenswrapper[4744]: I0106 15:31:26.309811 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" podStartSLOduration=1.801542291 podStartE2EDuration="2.30978609s" podCreationTimestamp="2026-01-06 15:31:24 +0000 UTC" firstStartedPulling="2026-01-06 15:31:25.346801379 +0000 UTC m=+3281.974267707" lastFinishedPulling="2026-01-06 15:31:25.855045188 +0000 UTC m=+3282.482511506" observedRunningTime="2026-01-06 15:31:26.29893728 +0000 UTC m=+3282.926403638" watchObservedRunningTime="2026-01-06 15:31:26.30978609 +0000 UTC m=+3282.937252428" Jan 06 15:31:34 crc kubenswrapper[4744]: I0106 15:31:34.902109 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:34 crc kubenswrapper[4744]: I0106 15:31:34.969124 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:35 crc kubenswrapper[4744]: I0106 15:31:35.659487 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lc5t7"] Jan 06 15:31:36 crc kubenswrapper[4744]: I0106 15:31:36.401576 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lc5t7" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="registry-server" containerID="cri-o://95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2" gracePeriod=2 Jan 06 15:31:36 crc kubenswrapper[4744]: I0106 15:31:36.954633 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.109089 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgbw7\" (UniqueName: \"kubernetes.io/projected/91efe524-a251-4191-84f1-a611aec44891-kube-api-access-fgbw7\") pod \"91efe524-a251-4191-84f1-a611aec44891\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.109438 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-utilities\") pod \"91efe524-a251-4191-84f1-a611aec44891\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.109785 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-catalog-content\") pod \"91efe524-a251-4191-84f1-a611aec44891\" (UID: \"91efe524-a251-4191-84f1-a611aec44891\") " Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.110329 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-utilities" (OuterVolumeSpecName: "utilities") pod "91efe524-a251-4191-84f1-a611aec44891" (UID: "91efe524-a251-4191-84f1-a611aec44891"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.110491 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.115042 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91efe524-a251-4191-84f1-a611aec44891-kube-api-access-fgbw7" (OuterVolumeSpecName: "kube-api-access-fgbw7") pod "91efe524-a251-4191-84f1-a611aec44891" (UID: "91efe524-a251-4191-84f1-a611aec44891"). InnerVolumeSpecName "kube-api-access-fgbw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.212748 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgbw7\" (UniqueName: \"kubernetes.io/projected/91efe524-a251-4191-84f1-a611aec44891-kube-api-access-fgbw7\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.255413 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91efe524-a251-4191-84f1-a611aec44891" (UID: "91efe524-a251-4191-84f1-a611aec44891"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.314982 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91efe524-a251-4191-84f1-a611aec44891-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.413294 4744 generic.go:334] "Generic (PLEG): container finished" podID="91efe524-a251-4191-84f1-a611aec44891" containerID="95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2" exitCode=0 Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.413337 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc5t7" event={"ID":"91efe524-a251-4191-84f1-a611aec44891","Type":"ContainerDied","Data":"95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2"} Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.413381 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lc5t7" event={"ID":"91efe524-a251-4191-84f1-a611aec44891","Type":"ContainerDied","Data":"1930b1c893ba066e1846d13b2f399f95718febbdacbbe84de73689de18ac0539"} Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.413388 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lc5t7" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.413401 4744 scope.go:117] "RemoveContainer" containerID="95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.435353 4744 scope.go:117] "RemoveContainer" containerID="312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.458248 4744 scope.go:117] "RemoveContainer" containerID="f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.475850 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lc5t7"] Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.489321 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lc5t7"] Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.529985 4744 scope.go:117] "RemoveContainer" containerID="95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2" Jan 06 15:31:37 crc kubenswrapper[4744]: E0106 15:31:37.530403 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2\": container with ID starting with 95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2 not found: ID does not exist" containerID="95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.530437 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2"} err="failed to get container status \"95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2\": rpc error: code = NotFound desc = could not find container \"95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2\": container with ID starting with 95a4adfed89950946cd96a0f14857bad9405b2be3e00b3962e6e1c31d232fea2 not found: ID does not exist" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.530458 4744 scope.go:117] "RemoveContainer" containerID="312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251" Jan 06 15:31:37 crc kubenswrapper[4744]: E0106 15:31:37.530871 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251\": container with ID starting with 312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251 not found: ID does not exist" containerID="312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.530913 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251"} err="failed to get container status \"312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251\": rpc error: code = NotFound desc = could not find container \"312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251\": container with ID starting with 312717a86e90c5bd914909cf458053c057ec56f0ece85df89f9de288d70a6251 not found: ID does not exist" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.530942 4744 scope.go:117] "RemoveContainer" containerID="f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4" Jan 06 15:31:37 crc kubenswrapper[4744]: E0106 15:31:37.531279 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4\": container with ID starting with f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4 not found: ID does not exist" containerID="f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.531305 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4"} err="failed to get container status \"f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4\": rpc error: code = NotFound desc = could not find container \"f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4\": container with ID starting with f8cbf0a2e9af298ddf04a16aedeb99f59eaab26e37ce7ffa0381840dc90ad9c4 not found: ID does not exist" Jan 06 15:31:37 crc kubenswrapper[4744]: I0106 15:31:37.727657 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91efe524-a251-4191-84f1-a611aec44891" path="/var/lib/kubelet/pods/91efe524-a251-4191-84f1-a611aec44891/volumes" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.111280 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5xt6n"] Jan 06 15:31:49 crc kubenswrapper[4744]: E0106 15:31:49.113422 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="registry-server" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.113536 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="registry-server" Jan 06 15:31:49 crc kubenswrapper[4744]: E0106 15:31:49.113627 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="extract-content" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.113714 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="extract-content" Jan 06 15:31:49 crc kubenswrapper[4744]: E0106 15:31:49.113821 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="extract-utilities" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.113994 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="extract-utilities" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.114416 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="91efe524-a251-4191-84f1-a611aec44891" containerName="registry-server" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.116707 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.153501 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xt6n"] Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.240609 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-catalog-content\") pod \"community-operators-5xt6n\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.240690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-utilities\") pod \"community-operators-5xt6n\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.240761 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvcpj\" (UniqueName: \"kubernetes.io/projected/3caa58b4-13a2-4324-8679-e42a1e15bbbf-kube-api-access-pvcpj\") pod \"community-operators-5xt6n\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.343270 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-catalog-content\") pod \"community-operators-5xt6n\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.343636 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-utilities\") pod \"community-operators-5xt6n\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.343700 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvcpj\" (UniqueName: \"kubernetes.io/projected/3caa58b4-13a2-4324-8679-e42a1e15bbbf-kube-api-access-pvcpj\") pod \"community-operators-5xt6n\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.343833 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-catalog-content\") pod \"community-operators-5xt6n\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.344190 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-utilities\") pod \"community-operators-5xt6n\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.365758 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvcpj\" (UniqueName: \"kubernetes.io/projected/3caa58b4-13a2-4324-8679-e42a1e15bbbf-kube-api-access-pvcpj\") pod \"community-operators-5xt6n\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:49 crc kubenswrapper[4744]: I0106 15:31:49.444016 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:50 crc kubenswrapper[4744]: I0106 15:31:50.112833 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xt6n"] Jan 06 15:31:50 crc kubenswrapper[4744]: I0106 15:31:50.586178 4744 generic.go:334] "Generic (PLEG): container finished" podID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerID="63d69d65e9fdfcc5c7fb59b7388aea2296671164dc958380189c17d987ca1225" exitCode=0 Jan 06 15:31:50 crc kubenswrapper[4744]: I0106 15:31:50.586451 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xt6n" event={"ID":"3caa58b4-13a2-4324-8679-e42a1e15bbbf","Type":"ContainerDied","Data":"63d69d65e9fdfcc5c7fb59b7388aea2296671164dc958380189c17d987ca1225"} Jan 06 15:31:50 crc kubenswrapper[4744]: I0106 15:31:50.586477 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xt6n" event={"ID":"3caa58b4-13a2-4324-8679-e42a1e15bbbf","Type":"ContainerStarted","Data":"ccb7b4ffd4964d15fc673513e13e955511e07160945fcc0717b7dc88bcb26ac8"} Jan 06 15:31:51 crc kubenswrapper[4744]: I0106 15:31:51.596753 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xt6n" event={"ID":"3caa58b4-13a2-4324-8679-e42a1e15bbbf","Type":"ContainerStarted","Data":"c82f49491ab3732a910743b6b6cd75e092d6e9686ea72b10b21b8cb57955ba2b"} Jan 06 15:31:52 crc kubenswrapper[4744]: I0106 15:31:52.610268 4744 generic.go:334] "Generic (PLEG): container finished" podID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerID="c82f49491ab3732a910743b6b6cd75e092d6e9686ea72b10b21b8cb57955ba2b" exitCode=0 Jan 06 15:31:52 crc kubenswrapper[4744]: I0106 15:31:52.610364 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xt6n" event={"ID":"3caa58b4-13a2-4324-8679-e42a1e15bbbf","Type":"ContainerDied","Data":"c82f49491ab3732a910743b6b6cd75e092d6e9686ea72b10b21b8cb57955ba2b"} Jan 06 15:31:53 crc kubenswrapper[4744]: I0106 15:31:53.624775 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xt6n" event={"ID":"3caa58b4-13a2-4324-8679-e42a1e15bbbf","Type":"ContainerStarted","Data":"705dd04537357e2690f16a817f5725b7f4c5075e387562342e1cb57ad70aed6c"} Jan 06 15:31:53 crc kubenswrapper[4744]: I0106 15:31:53.647705 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5xt6n" podStartSLOduration=2.075418737 podStartE2EDuration="4.647677817s" podCreationTimestamp="2026-01-06 15:31:49 +0000 UTC" firstStartedPulling="2026-01-06 15:31:50.588788575 +0000 UTC m=+3307.216254903" lastFinishedPulling="2026-01-06 15:31:53.161047655 +0000 UTC m=+3309.788513983" observedRunningTime="2026-01-06 15:31:53.641178354 +0000 UTC m=+3310.268644682" watchObservedRunningTime="2026-01-06 15:31:53.647677817 +0000 UTC m=+3310.275144135" Jan 06 15:31:59 crc kubenswrapper[4744]: I0106 15:31:59.444756 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:59 crc kubenswrapper[4744]: I0106 15:31:59.445617 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:59 crc kubenswrapper[4744]: I0106 15:31:59.514756 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:59 crc kubenswrapper[4744]: I0106 15:31:59.790257 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:31:59 crc kubenswrapper[4744]: I0106 15:31:59.848524 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xt6n"] Jan 06 15:32:01 crc kubenswrapper[4744]: I0106 15:32:01.746563 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5xt6n" podUID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerName="registry-server" containerID="cri-o://705dd04537357e2690f16a817f5725b7f4c5075e387562342e1cb57ad70aed6c" gracePeriod=2 Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.758409 4744 generic.go:334] "Generic (PLEG): container finished" podID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerID="705dd04537357e2690f16a817f5725b7f4c5075e387562342e1cb57ad70aed6c" exitCode=0 Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.758485 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xt6n" event={"ID":"3caa58b4-13a2-4324-8679-e42a1e15bbbf","Type":"ContainerDied","Data":"705dd04537357e2690f16a817f5725b7f4c5075e387562342e1cb57ad70aed6c"} Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.758880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xt6n" event={"ID":"3caa58b4-13a2-4324-8679-e42a1e15bbbf","Type":"ContainerDied","Data":"ccb7b4ffd4964d15fc673513e13e955511e07160945fcc0717b7dc88bcb26ac8"} Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.758897 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccb7b4ffd4964d15fc673513e13e955511e07160945fcc0717b7dc88bcb26ac8" Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.834905 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.892245 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-catalog-content\") pod \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.892368 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvcpj\" (UniqueName: \"kubernetes.io/projected/3caa58b4-13a2-4324-8679-e42a1e15bbbf-kube-api-access-pvcpj\") pod \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.892595 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-utilities\") pod \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\" (UID: \"3caa58b4-13a2-4324-8679-e42a1e15bbbf\") " Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.894369 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-utilities" (OuterVolumeSpecName: "utilities") pod "3caa58b4-13a2-4324-8679-e42a1e15bbbf" (UID: "3caa58b4-13a2-4324-8679-e42a1e15bbbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.907821 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3caa58b4-13a2-4324-8679-e42a1e15bbbf-kube-api-access-pvcpj" (OuterVolumeSpecName: "kube-api-access-pvcpj") pod "3caa58b4-13a2-4324-8679-e42a1e15bbbf" (UID: "3caa58b4-13a2-4324-8679-e42a1e15bbbf"). InnerVolumeSpecName "kube-api-access-pvcpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.948254 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3caa58b4-13a2-4324-8679-e42a1e15bbbf" (UID: "3caa58b4-13a2-4324-8679-e42a1e15bbbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.995329 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.995368 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3caa58b4-13a2-4324-8679-e42a1e15bbbf-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:32:02 crc kubenswrapper[4744]: I0106 15:32:02.995378 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvcpj\" (UniqueName: \"kubernetes.io/projected/3caa58b4-13a2-4324-8679-e42a1e15bbbf-kube-api-access-pvcpj\") on node \"crc\" DevicePath \"\"" Jan 06 15:32:03 crc kubenswrapper[4744]: I0106 15:32:03.769232 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xt6n" Jan 06 15:32:03 crc kubenswrapper[4744]: I0106 15:32:03.804113 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xt6n"] Jan 06 15:32:03 crc kubenswrapper[4744]: I0106 15:32:03.816116 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5xt6n"] Jan 06 15:32:05 crc kubenswrapper[4744]: I0106 15:32:05.733586 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" path="/var/lib/kubelet/pods/3caa58b4-13a2-4324-8679-e42a1e15bbbf/volumes" Jan 06 15:32:44 crc kubenswrapper[4744]: I0106 15:32:44.423400 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:32:44 crc kubenswrapper[4744]: I0106 15:32:44.423904 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:33:14 crc kubenswrapper[4744]: I0106 15:33:14.423610 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:33:14 crc kubenswrapper[4744]: I0106 15:33:14.424212 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:33:44 crc kubenswrapper[4744]: I0106 15:33:44.423448 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:33:44 crc kubenswrapper[4744]: I0106 15:33:44.424054 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:33:44 crc kubenswrapper[4744]: I0106 15:33:44.424109 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:33:44 crc kubenswrapper[4744]: I0106 15:33:44.425106 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"38d5508f3beb00f9bb486f7d949acf4397056b3e1d199506d1c8635be557972c"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:33:44 crc kubenswrapper[4744]: I0106 15:33:44.425295 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://38d5508f3beb00f9bb486f7d949acf4397056b3e1d199506d1c8635be557972c" gracePeriod=600 Jan 06 15:33:44 crc kubenswrapper[4744]: E0106 15:33:44.652782 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1639b17c_035d_4955_a3b6_ac33889e1c89.slice/crio-38d5508f3beb00f9bb486f7d949acf4397056b3e1d199506d1c8635be557972c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1639b17c_035d_4955_a3b6_ac33889e1c89.slice/crio-conmon-38d5508f3beb00f9bb486f7d949acf4397056b3e1d199506d1c8635be557972c.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:33:45 crc kubenswrapper[4744]: I0106 15:33:45.151343 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="38d5508f3beb00f9bb486f7d949acf4397056b3e1d199506d1c8635be557972c" exitCode=0 Jan 06 15:33:45 crc kubenswrapper[4744]: I0106 15:33:45.153758 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"38d5508f3beb00f9bb486f7d949acf4397056b3e1d199506d1c8635be557972c"} Jan 06 15:33:45 crc kubenswrapper[4744]: I0106 15:33:45.153897 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d"} Jan 06 15:33:45 crc kubenswrapper[4744]: I0106 15:33:45.158289 4744 scope.go:117] "RemoveContainer" containerID="8d06abe3b08d075b413c541da3c11b218231ba519e5b89b920ba02e6a5ad7afb" Jan 06 15:33:52 crc kubenswrapper[4744]: I0106 15:33:52.250452 4744 generic.go:334] "Generic (PLEG): container finished" podID="5216d2ec-d2e3-4f91-97d5-5839eaeb173d" containerID="0533b8addfaf4f5d766d5949829f6c13618dfe1a470614093b60085c9b5d1a38" exitCode=0 Jan 06 15:33:52 crc kubenswrapper[4744]: I0106 15:33:52.250522 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" event={"ID":"5216d2ec-d2e3-4f91-97d5-5839eaeb173d","Type":"ContainerDied","Data":"0533b8addfaf4f5d766d5949829f6c13618dfe1a470614093b60085c9b5d1a38"} Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.754866 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.804581 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-2\") pod \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.804639 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-1\") pod \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.804696 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-inventory\") pod \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.804782 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-telemetry-power-monitoring-combined-ca-bundle\") pod \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.804891 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dcvs\" (UniqueName: \"kubernetes.io/projected/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-kube-api-access-9dcvs\") pod \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.805027 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ssh-key-openstack-edpm-ipam\") pod \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.805072 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-0\") pod \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\" (UID: \"5216d2ec-d2e3-4f91-97d5-5839eaeb173d\") " Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.811246 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-kube-api-access-9dcvs" (OuterVolumeSpecName: "kube-api-access-9dcvs") pod "5216d2ec-d2e3-4f91-97d5-5839eaeb173d" (UID: "5216d2ec-d2e3-4f91-97d5-5839eaeb173d"). InnerVolumeSpecName "kube-api-access-9dcvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.812295 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "5216d2ec-d2e3-4f91-97d5-5839eaeb173d" (UID: "5216d2ec-d2e3-4f91-97d5-5839eaeb173d"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.840211 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "5216d2ec-d2e3-4f91-97d5-5839eaeb173d" (UID: "5216d2ec-d2e3-4f91-97d5-5839eaeb173d"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.849888 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-inventory" (OuterVolumeSpecName: "inventory") pod "5216d2ec-d2e3-4f91-97d5-5839eaeb173d" (UID: "5216d2ec-d2e3-4f91-97d5-5839eaeb173d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.853636 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "5216d2ec-d2e3-4f91-97d5-5839eaeb173d" (UID: "5216d2ec-d2e3-4f91-97d5-5839eaeb173d"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.859487 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5216d2ec-d2e3-4f91-97d5-5839eaeb173d" (UID: "5216d2ec-d2e3-4f91-97d5-5839eaeb173d"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.861328 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "5216d2ec-d2e3-4f91-97d5-5839eaeb173d" (UID: "5216d2ec-d2e3-4f91-97d5-5839eaeb173d"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.908561 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.908595 4744 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.908611 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dcvs\" (UniqueName: \"kubernetes.io/projected/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-kube-api-access-9dcvs\") on node \"crc\" DevicePath \"\"" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.908627 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.908640 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.908654 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Jan 06 15:33:53 crc kubenswrapper[4744]: I0106 15:33:53.908668 4744 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/5216d2ec-d2e3-4f91-97d5-5839eaeb173d-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.278313 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" event={"ID":"5216d2ec-d2e3-4f91-97d5-5839eaeb173d","Type":"ContainerDied","Data":"b0fd8d7d7dbc8758425a04ebce9bbb76bd2e6d59f43c13c54c109ec308ccfb3b"} Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.278585 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0fd8d7d7dbc8758425a04ebce9bbb76bd2e6d59f43c13c54c109ec308ccfb3b" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.278405 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.413310 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w"] Jan 06 15:33:54 crc kubenswrapper[4744]: E0106 15:33:54.414205 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5216d2ec-d2e3-4f91-97d5-5839eaeb173d" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.414315 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5216d2ec-d2e3-4f91-97d5-5839eaeb173d" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Jan 06 15:33:54 crc kubenswrapper[4744]: E0106 15:33:54.414391 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerName="extract-utilities" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.414553 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerName="extract-utilities" Jan 06 15:33:54 crc kubenswrapper[4744]: E0106 15:33:54.414639 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerName="extract-content" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.414723 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerName="extract-content" Jan 06 15:33:54 crc kubenswrapper[4744]: E0106 15:33:54.414847 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerName="registry-server" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.414923 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerName="registry-server" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.415330 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5216d2ec-d2e3-4f91-97d5-5839eaeb173d" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.415445 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3caa58b4-13a2-4324-8679-e42a1e15bbbf" containerName="registry-server" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.416582 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.420005 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-4rjk7" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.420341 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.426476 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.426556 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.426875 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.428877 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w"] Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.531273 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.531353 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bplgf\" (UniqueName: \"kubernetes.io/projected/362fd3ee-198d-4233-b757-8db8895dc05b-kube-api-access-bplgf\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.531414 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.531516 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.531596 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-ssh-key-openstack-edpm-ipam\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.634057 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.635116 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-ssh-key-openstack-edpm-ipam\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.635691 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.635812 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bplgf\" (UniqueName: \"kubernetes.io/projected/362fd3ee-198d-4233-b757-8db8895dc05b-kube-api-access-bplgf\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.635864 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.639418 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.639947 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.640582 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-ssh-key-openstack-edpm-ipam\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.641930 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.653756 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bplgf\" (UniqueName: \"kubernetes.io/projected/362fd3ee-198d-4233-b757-8db8895dc05b-kube-api-access-bplgf\") pod \"logging-edpm-deployment-openstack-edpm-ipam-pp45w\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:54 crc kubenswrapper[4744]: I0106 15:33:54.736638 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:33:55 crc kubenswrapper[4744]: I0106 15:33:55.335953 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w"] Jan 06 15:33:56 crc kubenswrapper[4744]: I0106 15:33:56.305750 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" event={"ID":"362fd3ee-198d-4233-b757-8db8895dc05b","Type":"ContainerStarted","Data":"1b4e928e343c8c3622612de1c71c37d9c1475b04045bfea9fcaacb0c17a6af59"} Jan 06 15:33:56 crc kubenswrapper[4744]: I0106 15:33:56.306462 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" event={"ID":"362fd3ee-198d-4233-b757-8db8895dc05b","Type":"ContainerStarted","Data":"e34cf6fe526b44e71ebed711a89965cc79f49d482cb48a30ca27ac4b69781f13"} Jan 06 15:33:56 crc kubenswrapper[4744]: I0106 15:33:56.333548 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" podStartSLOduration=1.889360446 podStartE2EDuration="2.333527805s" podCreationTimestamp="2026-01-06 15:33:54 +0000 UTC" firstStartedPulling="2026-01-06 15:33:55.334148454 +0000 UTC m=+3431.961614802" lastFinishedPulling="2026-01-06 15:33:55.778315803 +0000 UTC m=+3432.405782161" observedRunningTime="2026-01-06 15:33:56.328845991 +0000 UTC m=+3432.956312309" watchObservedRunningTime="2026-01-06 15:33:56.333527805 +0000 UTC m=+3432.960994123" Jan 06 15:34:12 crc kubenswrapper[4744]: I0106 15:34:12.716867 4744 generic.go:334] "Generic (PLEG): container finished" podID="362fd3ee-198d-4233-b757-8db8895dc05b" containerID="1b4e928e343c8c3622612de1c71c37d9c1475b04045bfea9fcaacb0c17a6af59" exitCode=0 Jan 06 15:34:12 crc kubenswrapper[4744]: I0106 15:34:12.716970 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" event={"ID":"362fd3ee-198d-4233-b757-8db8895dc05b","Type":"ContainerDied","Data":"1b4e928e343c8c3622612de1c71c37d9c1475b04045bfea9fcaacb0c17a6af59"} Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.250020 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.336612 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-ssh-key-openstack-edpm-ipam\") pod \"362fd3ee-198d-4233-b757-8db8895dc05b\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.336704 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-inventory\") pod \"362fd3ee-198d-4233-b757-8db8895dc05b\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.336989 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-0\") pod \"362fd3ee-198d-4233-b757-8db8895dc05b\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.337063 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-1\") pod \"362fd3ee-198d-4233-b757-8db8895dc05b\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.337364 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bplgf\" (UniqueName: \"kubernetes.io/projected/362fd3ee-198d-4233-b757-8db8895dc05b-kube-api-access-bplgf\") pod \"362fd3ee-198d-4233-b757-8db8895dc05b\" (UID: \"362fd3ee-198d-4233-b757-8db8895dc05b\") " Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.347404 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/362fd3ee-198d-4233-b757-8db8895dc05b-kube-api-access-bplgf" (OuterVolumeSpecName: "kube-api-access-bplgf") pod "362fd3ee-198d-4233-b757-8db8895dc05b" (UID: "362fd3ee-198d-4233-b757-8db8895dc05b"). InnerVolumeSpecName "kube-api-access-bplgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.380301 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "362fd3ee-198d-4233-b757-8db8895dc05b" (UID: "362fd3ee-198d-4233-b757-8db8895dc05b"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.381726 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "362fd3ee-198d-4233-b757-8db8895dc05b" (UID: "362fd3ee-198d-4233-b757-8db8895dc05b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.389669 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-inventory" (OuterVolumeSpecName: "inventory") pod "362fd3ee-198d-4233-b757-8db8895dc05b" (UID: "362fd3ee-198d-4233-b757-8db8895dc05b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.412509 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "362fd3ee-198d-4233-b757-8db8895dc05b" (UID: "362fd3ee-198d-4233-b757-8db8895dc05b"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.442551 4744 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.442597 4744 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.442610 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bplgf\" (UniqueName: \"kubernetes.io/projected/362fd3ee-198d-4233-b757-8db8895dc05b-kube-api-access-bplgf\") on node \"crc\" DevicePath \"\"" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.442620 4744 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.442631 4744 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/362fd3ee-198d-4233-b757-8db8895dc05b-inventory\") on node \"crc\" DevicePath \"\"" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.783536 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" event={"ID":"362fd3ee-198d-4233-b757-8db8895dc05b","Type":"ContainerDied","Data":"e34cf6fe526b44e71ebed711a89965cc79f49d482cb48a30ca27ac4b69781f13"} Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.783599 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e34cf6fe526b44e71ebed711a89965cc79f49d482cb48a30ca27ac4b69781f13" Jan 06 15:34:14 crc kubenswrapper[4744]: I0106 15:34:14.783721 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-pp45w" Jan 06 15:35:44 crc kubenswrapper[4744]: I0106 15:35:44.423895 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:35:44 crc kubenswrapper[4744]: I0106 15:35:44.424401 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:36:14 crc kubenswrapper[4744]: I0106 15:36:14.423580 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:36:14 crc kubenswrapper[4744]: I0106 15:36:14.424259 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:36:44 crc kubenswrapper[4744]: I0106 15:36:44.423976 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:36:44 crc kubenswrapper[4744]: I0106 15:36:44.424612 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:36:44 crc kubenswrapper[4744]: I0106 15:36:44.424677 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:36:44 crc kubenswrapper[4744]: I0106 15:36:44.425956 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:36:44 crc kubenswrapper[4744]: I0106 15:36:44.426052 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" gracePeriod=600 Jan 06 15:36:44 crc kubenswrapper[4744]: E0106 15:36:44.562211 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:36:44 crc kubenswrapper[4744]: I0106 15:36:44.625976 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" exitCode=0 Jan 06 15:36:44 crc kubenswrapper[4744]: I0106 15:36:44.626051 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d"} Jan 06 15:36:44 crc kubenswrapper[4744]: I0106 15:36:44.626292 4744 scope.go:117] "RemoveContainer" containerID="38d5508f3beb00f9bb486f7d949acf4397056b3e1d199506d1c8635be557972c" Jan 06 15:36:44 crc kubenswrapper[4744]: I0106 15:36:44.627045 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:36:44 crc kubenswrapper[4744]: E0106 15:36:44.627328 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:36:56 crc kubenswrapper[4744]: I0106 15:36:56.713924 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:36:56 crc kubenswrapper[4744]: E0106 15:36:56.715226 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:37:07 crc kubenswrapper[4744]: I0106 15:37:07.711083 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:37:07 crc kubenswrapper[4744]: E0106 15:37:07.711972 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:37:18 crc kubenswrapper[4744]: I0106 15:37:18.711293 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:37:18 crc kubenswrapper[4744]: E0106 15:37:18.712220 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:37:30 crc kubenswrapper[4744]: I0106 15:37:30.711641 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:37:30 crc kubenswrapper[4744]: E0106 15:37:30.712270 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:37:44 crc kubenswrapper[4744]: I0106 15:37:44.711615 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:37:44 crc kubenswrapper[4744]: E0106 15:37:44.712432 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:37:59 crc kubenswrapper[4744]: I0106 15:37:59.712190 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:37:59 crc kubenswrapper[4744]: E0106 15:37:59.713232 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:38:11 crc kubenswrapper[4744]: I0106 15:38:11.711770 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:38:11 crc kubenswrapper[4744]: E0106 15:38:11.713030 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:38:12 crc kubenswrapper[4744]: I0106 15:38:12.795036 4744 scope.go:117] "RemoveContainer" containerID="63d69d65e9fdfcc5c7fb59b7388aea2296671164dc958380189c17d987ca1225" Jan 06 15:38:12 crc kubenswrapper[4744]: I0106 15:38:12.835242 4744 scope.go:117] "RemoveContainer" containerID="705dd04537357e2690f16a817f5725b7f4c5075e387562342e1cb57ad70aed6c" Jan 06 15:38:12 crc kubenswrapper[4744]: I0106 15:38:12.912630 4744 scope.go:117] "RemoveContainer" containerID="c82f49491ab3732a910743b6b6cd75e092d6e9686ea72b10b21b8cb57955ba2b" Jan 06 15:38:25 crc kubenswrapper[4744]: I0106 15:38:25.712184 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:38:25 crc kubenswrapper[4744]: E0106 15:38:25.713208 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:38:37 crc kubenswrapper[4744]: I0106 15:38:37.711383 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:38:37 crc kubenswrapper[4744]: E0106 15:38:37.712445 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:38:48 crc kubenswrapper[4744]: I0106 15:38:48.711323 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:38:48 crc kubenswrapper[4744]: E0106 15:38:48.712231 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.719879 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dp2rv"] Jan 06 15:38:56 crc kubenswrapper[4744]: E0106 15:38:56.721310 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="362fd3ee-198d-4233-b757-8db8895dc05b" containerName="logging-edpm-deployment-openstack-edpm-ipam" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.721337 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="362fd3ee-198d-4233-b757-8db8895dc05b" containerName="logging-edpm-deployment-openstack-edpm-ipam" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.721793 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="362fd3ee-198d-4233-b757-8db8895dc05b" containerName="logging-edpm-deployment-openstack-edpm-ipam" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.724901 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.750002 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp2rv"] Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.838511 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-572zc\" (UniqueName: \"kubernetes.io/projected/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-kube-api-access-572zc\") pod \"redhat-marketplace-dp2rv\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.838865 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-catalog-content\") pod \"redhat-marketplace-dp2rv\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.839138 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-utilities\") pod \"redhat-marketplace-dp2rv\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.942323 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-utilities\") pod \"redhat-marketplace-dp2rv\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.942736 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-572zc\" (UniqueName: \"kubernetes.io/projected/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-kube-api-access-572zc\") pod \"redhat-marketplace-dp2rv\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.942871 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-catalog-content\") pod \"redhat-marketplace-dp2rv\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.942957 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-utilities\") pod \"redhat-marketplace-dp2rv\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.943381 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-catalog-content\") pod \"redhat-marketplace-dp2rv\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:56 crc kubenswrapper[4744]: I0106 15:38:56.962320 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-572zc\" (UniqueName: \"kubernetes.io/projected/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-kube-api-access-572zc\") pod \"redhat-marketplace-dp2rv\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:57 crc kubenswrapper[4744]: I0106 15:38:57.061649 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:38:57 crc kubenswrapper[4744]: I0106 15:38:57.538347 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp2rv"] Jan 06 15:38:58 crc kubenswrapper[4744]: I0106 15:38:58.341853 4744 generic.go:334] "Generic (PLEG): container finished" podID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerID="002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff" exitCode=0 Jan 06 15:38:58 crc kubenswrapper[4744]: I0106 15:38:58.341929 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp2rv" event={"ID":"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7","Type":"ContainerDied","Data":"002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff"} Jan 06 15:38:58 crc kubenswrapper[4744]: I0106 15:38:58.342245 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp2rv" event={"ID":"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7","Type":"ContainerStarted","Data":"d310a7267e79bfbbb19cea6ca783fc9bc3bbf7270aa8eee120472dbe9f3da521"} Jan 06 15:38:58 crc kubenswrapper[4744]: I0106 15:38:58.345536 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 15:38:59 crc kubenswrapper[4744]: I0106 15:38:59.355538 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp2rv" event={"ID":"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7","Type":"ContainerStarted","Data":"9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f"} Jan 06 15:39:00 crc kubenswrapper[4744]: I0106 15:39:00.369120 4744 generic.go:334] "Generic (PLEG): container finished" podID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerID="9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f" exitCode=0 Jan 06 15:39:00 crc kubenswrapper[4744]: I0106 15:39:00.369292 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp2rv" event={"ID":"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7","Type":"ContainerDied","Data":"9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f"} Jan 06 15:39:01 crc kubenswrapper[4744]: I0106 15:39:01.381880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp2rv" event={"ID":"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7","Type":"ContainerStarted","Data":"86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34"} Jan 06 15:39:01 crc kubenswrapper[4744]: I0106 15:39:01.403669 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dp2rv" podStartSLOduration=2.847686734 podStartE2EDuration="5.403649451s" podCreationTimestamp="2026-01-06 15:38:56 +0000 UTC" firstStartedPulling="2026-01-06 15:38:58.345090373 +0000 UTC m=+3734.972556731" lastFinishedPulling="2026-01-06 15:39:00.90105312 +0000 UTC m=+3737.528519448" observedRunningTime="2026-01-06 15:39:01.397301342 +0000 UTC m=+3738.024767670" watchObservedRunningTime="2026-01-06 15:39:01.403649451 +0000 UTC m=+3738.031115779" Jan 06 15:39:02 crc kubenswrapper[4744]: I0106 15:39:02.711153 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:39:02 crc kubenswrapper[4744]: E0106 15:39:02.711942 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:39:07 crc kubenswrapper[4744]: I0106 15:39:07.062312 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:39:07 crc kubenswrapper[4744]: I0106 15:39:07.063102 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:39:07 crc kubenswrapper[4744]: I0106 15:39:07.142357 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:39:07 crc kubenswrapper[4744]: I0106 15:39:07.531058 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:39:07 crc kubenswrapper[4744]: I0106 15:39:07.598459 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp2rv"] Jan 06 15:39:09 crc kubenswrapper[4744]: I0106 15:39:09.478799 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dp2rv" podUID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerName="registry-server" containerID="cri-o://86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34" gracePeriod=2 Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.222583 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.408269 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-catalog-content\") pod \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.408483 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-572zc\" (UniqueName: \"kubernetes.io/projected/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-kube-api-access-572zc\") pod \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.408537 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-utilities\") pod \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\" (UID: \"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7\") " Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.411727 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-utilities" (OuterVolumeSpecName: "utilities") pod "6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" (UID: "6bf56e52-cbe2-435a-a3f6-6e96b102e2f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.437556 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-kube-api-access-572zc" (OuterVolumeSpecName: "kube-api-access-572zc") pod "6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" (UID: "6bf56e52-cbe2-435a-a3f6-6e96b102e2f7"). InnerVolumeSpecName "kube-api-access-572zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.452422 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" (UID: "6bf56e52-cbe2-435a-a3f6-6e96b102e2f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.492216 4744 generic.go:334] "Generic (PLEG): container finished" podID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerID="86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34" exitCode=0 Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.492266 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp2rv" event={"ID":"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7","Type":"ContainerDied","Data":"86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34"} Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.492294 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dp2rv" event={"ID":"6bf56e52-cbe2-435a-a3f6-6e96b102e2f7","Type":"ContainerDied","Data":"d310a7267e79bfbbb19cea6ca783fc9bc3bbf7270aa8eee120472dbe9f3da521"} Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.492313 4744 scope.go:117] "RemoveContainer" containerID="86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.492466 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dp2rv" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.513585 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-572zc\" (UniqueName: \"kubernetes.io/projected/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-kube-api-access-572zc\") on node \"crc\" DevicePath \"\"" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.513626 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.513637 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.522149 4744 scope.go:117] "RemoveContainer" containerID="9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.539264 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp2rv"] Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.547648 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dp2rv"] Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.548390 4744 scope.go:117] "RemoveContainer" containerID="002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.620530 4744 scope.go:117] "RemoveContainer" containerID="86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34" Jan 06 15:39:10 crc kubenswrapper[4744]: E0106 15:39:10.621081 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34\": container with ID starting with 86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34 not found: ID does not exist" containerID="86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.621128 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34"} err="failed to get container status \"86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34\": rpc error: code = NotFound desc = could not find container \"86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34\": container with ID starting with 86c120abe7ed03b8ed3aec6bb49ff50b7010906e2f5611bbabf1db590d924f34 not found: ID does not exist" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.621370 4744 scope.go:117] "RemoveContainer" containerID="9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f" Jan 06 15:39:10 crc kubenswrapper[4744]: E0106 15:39:10.621831 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f\": container with ID starting with 9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f not found: ID does not exist" containerID="9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.621865 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f"} err="failed to get container status \"9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f\": rpc error: code = NotFound desc = could not find container \"9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f\": container with ID starting with 9b0fdf4bb93dd3ed61e4a2555e6ac8b907e27432122f7f251bdc54e0c574cd1f not found: ID does not exist" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.621892 4744 scope.go:117] "RemoveContainer" containerID="002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff" Jan 06 15:39:10 crc kubenswrapper[4744]: E0106 15:39:10.622377 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff\": container with ID starting with 002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff not found: ID does not exist" containerID="002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff" Jan 06 15:39:10 crc kubenswrapper[4744]: I0106 15:39:10.622424 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff"} err="failed to get container status \"002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff\": rpc error: code = NotFound desc = could not find container \"002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff\": container with ID starting with 002c0ecb6255da43fd1c1a2da57b1b5663bb99a5edef0fe5da87693368f586ff not found: ID does not exist" Jan 06 15:39:11 crc kubenswrapper[4744]: I0106 15:39:11.728846 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" path="/var/lib/kubelet/pods/6bf56e52-cbe2-435a-a3f6-6e96b102e2f7/volumes" Jan 06 15:39:15 crc kubenswrapper[4744]: I0106 15:39:15.711984 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:39:15 crc kubenswrapper[4744]: E0106 15:39:15.714190 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:39:26 crc kubenswrapper[4744]: I0106 15:39:26.714858 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:39:26 crc kubenswrapper[4744]: E0106 15:39:26.716519 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:39:38 crc kubenswrapper[4744]: I0106 15:39:38.712586 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:39:38 crc kubenswrapper[4744]: E0106 15:39:38.713409 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:39:52 crc kubenswrapper[4744]: I0106 15:39:52.711069 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:39:52 crc kubenswrapper[4744]: E0106 15:39:52.712214 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:40:06 crc kubenswrapper[4744]: I0106 15:40:06.711384 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:40:06 crc kubenswrapper[4744]: E0106 15:40:06.712397 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:40:20 crc kubenswrapper[4744]: I0106 15:40:20.711927 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:40:20 crc kubenswrapper[4744]: E0106 15:40:20.712946 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:40:33 crc kubenswrapper[4744]: I0106 15:40:33.724182 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:40:33 crc kubenswrapper[4744]: E0106 15:40:33.725438 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:40:46 crc kubenswrapper[4744]: I0106 15:40:46.711406 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:40:46 crc kubenswrapper[4744]: E0106 15:40:46.712690 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:40:59 crc kubenswrapper[4744]: I0106 15:40:59.711056 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:40:59 crc kubenswrapper[4744]: E0106 15:40:59.712366 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.276090 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bmqk4"] Jan 06 15:41:07 crc kubenswrapper[4744]: E0106 15:41:07.277276 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerName="extract-utilities" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.277290 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerName="extract-utilities" Jan 06 15:41:07 crc kubenswrapper[4744]: E0106 15:41:07.277301 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerName="extract-content" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.277307 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerName="extract-content" Jan 06 15:41:07 crc kubenswrapper[4744]: E0106 15:41:07.277338 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerName="registry-server" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.277347 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerName="registry-server" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.277612 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bf56e52-cbe2-435a-a3f6-6e96b102e2f7" containerName="registry-server" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.281648 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.296857 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bmqk4"] Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.357007 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-catalog-content\") pod \"certified-operators-bmqk4\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.357090 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcgv2\" (UniqueName: \"kubernetes.io/projected/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-kube-api-access-zcgv2\") pod \"certified-operators-bmqk4\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.357202 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-utilities\") pod \"certified-operators-bmqk4\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.459746 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcgv2\" (UniqueName: \"kubernetes.io/projected/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-kube-api-access-zcgv2\") pod \"certified-operators-bmqk4\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.459898 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-utilities\") pod \"certified-operators-bmqk4\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.460066 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-catalog-content\") pod \"certified-operators-bmqk4\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.460922 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-utilities\") pod \"certified-operators-bmqk4\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.460940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-catalog-content\") pod \"certified-operators-bmqk4\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.480727 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcgv2\" (UniqueName: \"kubernetes.io/projected/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-kube-api-access-zcgv2\") pod \"certified-operators-bmqk4\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:07 crc kubenswrapper[4744]: I0106 15:41:07.612417 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:08 crc kubenswrapper[4744]: I0106 15:41:08.212867 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bmqk4"] Jan 06 15:41:08 crc kubenswrapper[4744]: I0106 15:41:08.980759 4744 generic.go:334] "Generic (PLEG): container finished" podID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerID="de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc" exitCode=0 Jan 06 15:41:08 crc kubenswrapper[4744]: I0106 15:41:08.981129 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmqk4" event={"ID":"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc","Type":"ContainerDied","Data":"de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc"} Jan 06 15:41:08 crc kubenswrapper[4744]: I0106 15:41:08.981326 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmqk4" event={"ID":"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc","Type":"ContainerStarted","Data":"f8ec8b281887433cbd8196afdac0f0593ce306c5126480d7f2fa3756d7eab0cb"} Jan 06 15:41:09 crc kubenswrapper[4744]: I0106 15:41:09.996881 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmqk4" event={"ID":"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc","Type":"ContainerStarted","Data":"10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e"} Jan 06 15:41:12 crc kubenswrapper[4744]: I0106 15:41:12.021966 4744 generic.go:334] "Generic (PLEG): container finished" podID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerID="10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e" exitCode=0 Jan 06 15:41:12 crc kubenswrapper[4744]: I0106 15:41:12.022047 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmqk4" event={"ID":"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc","Type":"ContainerDied","Data":"10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e"} Jan 06 15:41:13 crc kubenswrapper[4744]: I0106 15:41:13.044769 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmqk4" event={"ID":"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc","Type":"ContainerStarted","Data":"5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f"} Jan 06 15:41:13 crc kubenswrapper[4744]: I0106 15:41:13.078710 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bmqk4" podStartSLOduration=2.615593998 podStartE2EDuration="6.078682031s" podCreationTimestamp="2026-01-06 15:41:07 +0000 UTC" firstStartedPulling="2026-01-06 15:41:08.984869456 +0000 UTC m=+3865.612335814" lastFinishedPulling="2026-01-06 15:41:12.447957529 +0000 UTC m=+3869.075423847" observedRunningTime="2026-01-06 15:41:13.066476536 +0000 UTC m=+3869.693942864" watchObservedRunningTime="2026-01-06 15:41:13.078682031 +0000 UTC m=+3869.706148349" Jan 06 15:41:14 crc kubenswrapper[4744]: I0106 15:41:14.712112 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:41:14 crc kubenswrapper[4744]: E0106 15:41:14.712909 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:41:17 crc kubenswrapper[4744]: I0106 15:41:17.613290 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:17 crc kubenswrapper[4744]: I0106 15:41:17.614087 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:17 crc kubenswrapper[4744]: I0106 15:41:17.666709 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:18 crc kubenswrapper[4744]: I0106 15:41:18.187040 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:18 crc kubenswrapper[4744]: I0106 15:41:18.245770 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bmqk4"] Jan 06 15:41:20 crc kubenswrapper[4744]: I0106 15:41:20.130092 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bmqk4" podUID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerName="registry-server" containerID="cri-o://5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f" gracePeriod=2 Jan 06 15:41:20 crc kubenswrapper[4744]: I0106 15:41:20.833945 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:20 crc kubenswrapper[4744]: I0106 15:41:20.972574 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-utilities\") pod \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " Jan 06 15:41:20 crc kubenswrapper[4744]: I0106 15:41:20.972738 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-catalog-content\") pod \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " Jan 06 15:41:20 crc kubenswrapper[4744]: I0106 15:41:20.973208 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcgv2\" (UniqueName: \"kubernetes.io/projected/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-kube-api-access-zcgv2\") pod \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\" (UID: \"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc\") " Jan 06 15:41:20 crc kubenswrapper[4744]: I0106 15:41:20.973842 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-utilities" (OuterVolumeSpecName: "utilities") pod "8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" (UID: "8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:41:20 crc kubenswrapper[4744]: I0106 15:41:20.975298 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:41:20 crc kubenswrapper[4744]: I0106 15:41:20.985024 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-kube-api-access-zcgv2" (OuterVolumeSpecName: "kube-api-access-zcgv2") pod "8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" (UID: "8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc"). InnerVolumeSpecName "kube-api-access-zcgv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.036116 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" (UID: "8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.078141 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcgv2\" (UniqueName: \"kubernetes.io/projected/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-kube-api-access-zcgv2\") on node \"crc\" DevicePath \"\"" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.078188 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.145271 4744 generic.go:334] "Generic (PLEG): container finished" podID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerID="5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f" exitCode=0 Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.145330 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmqk4" event={"ID":"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc","Type":"ContainerDied","Data":"5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f"} Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.145370 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bmqk4" event={"ID":"8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc","Type":"ContainerDied","Data":"f8ec8b281887433cbd8196afdac0f0593ce306c5126480d7f2fa3756d7eab0cb"} Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.145390 4744 scope.go:117] "RemoveContainer" containerID="5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.145414 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bmqk4" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.179135 4744 scope.go:117] "RemoveContainer" containerID="10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.213757 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bmqk4"] Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.227391 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bmqk4"] Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.238485 4744 scope.go:117] "RemoveContainer" containerID="de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.303522 4744 scope.go:117] "RemoveContainer" containerID="5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f" Jan 06 15:41:21 crc kubenswrapper[4744]: E0106 15:41:21.304276 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f\": container with ID starting with 5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f not found: ID does not exist" containerID="5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.304353 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f"} err="failed to get container status \"5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f\": rpc error: code = NotFound desc = could not find container \"5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f\": container with ID starting with 5773d6748cd16d81967f101d4c7f765517f96a8f2b19e043972c4b9aacfd127f not found: ID does not exist" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.304430 4744 scope.go:117] "RemoveContainer" containerID="10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e" Jan 06 15:41:21 crc kubenswrapper[4744]: E0106 15:41:21.305210 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e\": container with ID starting with 10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e not found: ID does not exist" containerID="10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.305262 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e"} err="failed to get container status \"10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e\": rpc error: code = NotFound desc = could not find container \"10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e\": container with ID starting with 10a8ddd43a3c19282c0e95e25a3c8d07288a2d21541e2cc6e8cc2523a8c0196e not found: ID does not exist" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.305287 4744 scope.go:117] "RemoveContainer" containerID="de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc" Jan 06 15:41:21 crc kubenswrapper[4744]: E0106 15:41:21.306109 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc\": container with ID starting with de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc not found: ID does not exist" containerID="de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.306256 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc"} err="failed to get container status \"de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc\": rpc error: code = NotFound desc = could not find container \"de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc\": container with ID starting with de2f5b60c5997aab6283da3b88cca42193f5dd4f5ec8da9f7fbcab6c9f467bbc not found: ID does not exist" Jan 06 15:41:21 crc kubenswrapper[4744]: I0106 15:41:21.749856 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" path="/var/lib/kubelet/pods/8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc/volumes" Jan 06 15:41:26 crc kubenswrapper[4744]: I0106 15:41:26.712325 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:41:26 crc kubenswrapper[4744]: E0106 15:41:26.713264 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:41:39 crc kubenswrapper[4744]: I0106 15:41:39.712073 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:41:39 crc kubenswrapper[4744]: E0106 15:41:39.713988 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:41:50 crc kubenswrapper[4744]: I0106 15:41:50.710836 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:41:51 crc kubenswrapper[4744]: I0106 15:41:51.550562 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"bc3d927402f59963ca59e24827db41e1cf066041e70d1875be675145db361865"} Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.706239 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rf899"] Jan 06 15:42:15 crc kubenswrapper[4744]: E0106 15:42:15.707322 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerName="extract-utilities" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.707337 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerName="extract-utilities" Jan 06 15:42:15 crc kubenswrapper[4744]: E0106 15:42:15.707359 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerName="extract-content" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.707367 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerName="extract-content" Jan 06 15:42:15 crc kubenswrapper[4744]: E0106 15:42:15.707389 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerName="registry-server" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.707398 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerName="registry-server" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.707678 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8cdc65-7c27-43d3-b9fe-7ee5c35cb2bc" containerName="registry-server" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.709757 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.745282 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rf899"] Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.792104 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-utilities\") pod \"redhat-operators-rf899\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.792401 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-catalog-content\") pod \"redhat-operators-rf899\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.792508 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkkmv\" (UniqueName: \"kubernetes.io/projected/b28635ea-49a5-4e03-a70d-54319e7ec3f7-kube-api-access-xkkmv\") pod \"redhat-operators-rf899\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.894929 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-catalog-content\") pod \"redhat-operators-rf899\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.895002 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkkmv\" (UniqueName: \"kubernetes.io/projected/b28635ea-49a5-4e03-a70d-54319e7ec3f7-kube-api-access-xkkmv\") pod \"redhat-operators-rf899\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.895231 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-utilities\") pod \"redhat-operators-rf899\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.895520 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-catalog-content\") pod \"redhat-operators-rf899\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.895676 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-utilities\") pod \"redhat-operators-rf899\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:15 crc kubenswrapper[4744]: I0106 15:42:15.933756 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkkmv\" (UniqueName: \"kubernetes.io/projected/b28635ea-49a5-4e03-a70d-54319e7ec3f7-kube-api-access-xkkmv\") pod \"redhat-operators-rf899\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:16 crc kubenswrapper[4744]: I0106 15:42:16.046937 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:16 crc kubenswrapper[4744]: I0106 15:42:16.591581 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rf899"] Jan 06 15:42:16 crc kubenswrapper[4744]: I0106 15:42:16.880860 4744 generic.go:334] "Generic (PLEG): container finished" podID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerID="803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6" exitCode=0 Jan 06 15:42:16 crc kubenswrapper[4744]: I0106 15:42:16.880933 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf899" event={"ID":"b28635ea-49a5-4e03-a70d-54319e7ec3f7","Type":"ContainerDied","Data":"803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6"} Jan 06 15:42:16 crc kubenswrapper[4744]: I0106 15:42:16.881519 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf899" event={"ID":"b28635ea-49a5-4e03-a70d-54319e7ec3f7","Type":"ContainerStarted","Data":"b644f3e207de9ae988d6dc746d7577ea888c360046cb902d41b5eecb1c931933"} Jan 06 15:42:17 crc kubenswrapper[4744]: I0106 15:42:17.893503 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf899" event={"ID":"b28635ea-49a5-4e03-a70d-54319e7ec3f7","Type":"ContainerStarted","Data":"c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17"} Jan 06 15:42:17 crc kubenswrapper[4744]: E0106 15:42:17.907486 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.22:57252->38.102.83.22:45165: write tcp 38.102.83.22:57252->38.102.83.22:45165: write: broken pipe Jan 06 15:42:21 crc kubenswrapper[4744]: I0106 15:42:21.943988 4744 generic.go:334] "Generic (PLEG): container finished" podID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerID="c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17" exitCode=0 Jan 06 15:42:21 crc kubenswrapper[4744]: I0106 15:42:21.944057 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf899" event={"ID":"b28635ea-49a5-4e03-a70d-54319e7ec3f7","Type":"ContainerDied","Data":"c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17"} Jan 06 15:42:22 crc kubenswrapper[4744]: I0106 15:42:22.963342 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf899" event={"ID":"b28635ea-49a5-4e03-a70d-54319e7ec3f7","Type":"ContainerStarted","Data":"4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0"} Jan 06 15:42:22 crc kubenswrapper[4744]: I0106 15:42:22.995712 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rf899" podStartSLOduration=2.486637465 podStartE2EDuration="7.995676613s" podCreationTimestamp="2026-01-06 15:42:15 +0000 UTC" firstStartedPulling="2026-01-06 15:42:16.882732157 +0000 UTC m=+3933.510198475" lastFinishedPulling="2026-01-06 15:42:22.391771265 +0000 UTC m=+3939.019237623" observedRunningTime="2026-01-06 15:42:22.983685244 +0000 UTC m=+3939.611151572" watchObservedRunningTime="2026-01-06 15:42:22.995676613 +0000 UTC m=+3939.623142951" Jan 06 15:42:26 crc kubenswrapper[4744]: I0106 15:42:26.048013 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:26 crc kubenswrapper[4744]: I0106 15:42:26.048544 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:27 crc kubenswrapper[4744]: I0106 15:42:27.140150 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rf899" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerName="registry-server" probeResult="failure" output=< Jan 06 15:42:27 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 15:42:27 crc kubenswrapper[4744]: > Jan 06 15:42:36 crc kubenswrapper[4744]: I0106 15:42:36.136610 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:36 crc kubenswrapper[4744]: I0106 15:42:36.283383 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:36 crc kubenswrapper[4744]: I0106 15:42:36.405620 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rf899"] Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.135517 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rf899" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerName="registry-server" containerID="cri-o://4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0" gracePeriod=2 Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.688535 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.774099 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-utilities\") pod \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.774186 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-catalog-content\") pod \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.774425 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkkmv\" (UniqueName: \"kubernetes.io/projected/b28635ea-49a5-4e03-a70d-54319e7ec3f7-kube-api-access-xkkmv\") pod \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\" (UID: \"b28635ea-49a5-4e03-a70d-54319e7ec3f7\") " Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.777972 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-utilities" (OuterVolumeSpecName: "utilities") pod "b28635ea-49a5-4e03-a70d-54319e7ec3f7" (UID: "b28635ea-49a5-4e03-a70d-54319e7ec3f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.797104 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b28635ea-49a5-4e03-a70d-54319e7ec3f7-kube-api-access-xkkmv" (OuterVolumeSpecName: "kube-api-access-xkkmv") pod "b28635ea-49a5-4e03-a70d-54319e7ec3f7" (UID: "b28635ea-49a5-4e03-a70d-54319e7ec3f7"). InnerVolumeSpecName "kube-api-access-xkkmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.878258 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkkmv\" (UniqueName: \"kubernetes.io/projected/b28635ea-49a5-4e03-a70d-54319e7ec3f7-kube-api-access-xkkmv\") on node \"crc\" DevicePath \"\"" Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.878296 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.906391 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b28635ea-49a5-4e03-a70d-54319e7ec3f7" (UID: "b28635ea-49a5-4e03-a70d-54319e7ec3f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:42:38 crc kubenswrapper[4744]: I0106 15:42:38.981294 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b28635ea-49a5-4e03-a70d-54319e7ec3f7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.150651 4744 generic.go:334] "Generic (PLEG): container finished" podID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerID="4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0" exitCode=0 Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.150733 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf899" event={"ID":"b28635ea-49a5-4e03-a70d-54319e7ec3f7","Type":"ContainerDied","Data":"4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0"} Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.151118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rf899" event={"ID":"b28635ea-49a5-4e03-a70d-54319e7ec3f7","Type":"ContainerDied","Data":"b644f3e207de9ae988d6dc746d7577ea888c360046cb902d41b5eecb1c931933"} Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.150745 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rf899" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.151233 4744 scope.go:117] "RemoveContainer" containerID="4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.185324 4744 scope.go:117] "RemoveContainer" containerID="c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.202122 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rf899"] Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.214845 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rf899"] Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.226124 4744 scope.go:117] "RemoveContainer" containerID="803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.294479 4744 scope.go:117] "RemoveContainer" containerID="4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0" Jan 06 15:42:39 crc kubenswrapper[4744]: E0106 15:42:39.294995 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0\": container with ID starting with 4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0 not found: ID does not exist" containerID="4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.295115 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0"} err="failed to get container status \"4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0\": rpc error: code = NotFound desc = could not find container \"4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0\": container with ID starting with 4275d27d6846a1ba203bd32c1e6fc33fdf6931745f8c993187d7bea0746e9dd0 not found: ID does not exist" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.295143 4744 scope.go:117] "RemoveContainer" containerID="c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17" Jan 06 15:42:39 crc kubenswrapper[4744]: E0106 15:42:39.295575 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17\": container with ID starting with c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17 not found: ID does not exist" containerID="c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.295602 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17"} err="failed to get container status \"c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17\": rpc error: code = NotFound desc = could not find container \"c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17\": container with ID starting with c21d9cbe594ef8cea1fc3fe6ec2da69b34e4e7773bafa70724308690df683a17 not found: ID does not exist" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.295624 4744 scope.go:117] "RemoveContainer" containerID="803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6" Jan 06 15:42:39 crc kubenswrapper[4744]: E0106 15:42:39.295875 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6\": container with ID starting with 803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6 not found: ID does not exist" containerID="803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.295908 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6"} err="failed to get container status \"803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6\": rpc error: code = NotFound desc = could not find container \"803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6\": container with ID starting with 803ecdffbffd9b06c863af86a26c55f7ccce650babd900807e9a2b5c9179bbf6 not found: ID does not exist" Jan 06 15:42:39 crc kubenswrapper[4744]: I0106 15:42:39.729544 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" path="/var/lib/kubelet/pods/b28635ea-49a5-4e03-a70d-54319e7ec3f7/volumes" Jan 06 15:44:00 crc kubenswrapper[4744]: E0106 15:44:00.347984 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.22:40878->38.102.83.22:45165: write tcp 38.102.83.22:40878->38.102.83.22:45165: write: connection reset by peer Jan 06 15:44:14 crc kubenswrapper[4744]: I0106 15:44:14.423820 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:44:14 crc kubenswrapper[4744]: I0106 15:44:14.424346 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:44:44 crc kubenswrapper[4744]: I0106 15:44:44.424437 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:44:44 crc kubenswrapper[4744]: I0106 15:44:44.424954 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.753516 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tvf6j"] Jan 06 15:44:45 crc kubenswrapper[4744]: E0106 15:44:45.754464 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerName="extract-content" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.754482 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerName="extract-content" Jan 06 15:44:45 crc kubenswrapper[4744]: E0106 15:44:45.754501 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerName="extract-utilities" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.754510 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerName="extract-utilities" Jan 06 15:44:45 crc kubenswrapper[4744]: E0106 15:44:45.754572 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerName="registry-server" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.754581 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerName="registry-server" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.754865 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="b28635ea-49a5-4e03-a70d-54319e7ec3f7" containerName="registry-server" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.761491 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.795867 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvf6j"] Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.849626 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92csz\" (UniqueName: \"kubernetes.io/projected/a8e774c8-20c3-44db-87d3-9431bc0c6e13-kube-api-access-92csz\") pod \"community-operators-tvf6j\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.849717 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-catalog-content\") pod \"community-operators-tvf6j\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.849909 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-utilities\") pod \"community-operators-tvf6j\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.952039 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-utilities\") pod \"community-operators-tvf6j\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.952446 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92csz\" (UniqueName: \"kubernetes.io/projected/a8e774c8-20c3-44db-87d3-9431bc0c6e13-kube-api-access-92csz\") pod \"community-operators-tvf6j\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.952547 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-utilities\") pod \"community-operators-tvf6j\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.952708 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-catalog-content\") pod \"community-operators-tvf6j\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.952948 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-catalog-content\") pod \"community-operators-tvf6j\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:45 crc kubenswrapper[4744]: I0106 15:44:45.972925 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92csz\" (UniqueName: \"kubernetes.io/projected/a8e774c8-20c3-44db-87d3-9431bc0c6e13-kube-api-access-92csz\") pod \"community-operators-tvf6j\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:46 crc kubenswrapper[4744]: I0106 15:44:46.092205 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:46 crc kubenswrapper[4744]: I0106 15:44:46.630328 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tvf6j"] Jan 06 15:44:47 crc kubenswrapper[4744]: I0106 15:44:47.823251 4744 generic.go:334] "Generic (PLEG): container finished" podID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerID="93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e" exitCode=0 Jan 06 15:44:47 crc kubenswrapper[4744]: I0106 15:44:47.823536 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvf6j" event={"ID":"a8e774c8-20c3-44db-87d3-9431bc0c6e13","Type":"ContainerDied","Data":"93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e"} Jan 06 15:44:47 crc kubenswrapper[4744]: I0106 15:44:47.823902 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvf6j" event={"ID":"a8e774c8-20c3-44db-87d3-9431bc0c6e13","Type":"ContainerStarted","Data":"343b66815b3fa96efdbe955352db7c2a8ec83707179555d5a02ab113ee3c7ce7"} Jan 06 15:44:47 crc kubenswrapper[4744]: I0106 15:44:47.825675 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 15:44:49 crc kubenswrapper[4744]: I0106 15:44:49.854395 4744 generic.go:334] "Generic (PLEG): container finished" podID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerID="02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69" exitCode=0 Jan 06 15:44:49 crc kubenswrapper[4744]: I0106 15:44:49.854477 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvf6j" event={"ID":"a8e774c8-20c3-44db-87d3-9431bc0c6e13","Type":"ContainerDied","Data":"02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69"} Jan 06 15:44:50 crc kubenswrapper[4744]: I0106 15:44:50.865766 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvf6j" event={"ID":"a8e774c8-20c3-44db-87d3-9431bc0c6e13","Type":"ContainerStarted","Data":"8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83"} Jan 06 15:44:50 crc kubenswrapper[4744]: I0106 15:44:50.896087 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tvf6j" podStartSLOduration=3.166527257 podStartE2EDuration="5.896067392s" podCreationTimestamp="2026-01-06 15:44:45 +0000 UTC" firstStartedPulling="2026-01-06 15:44:47.825234476 +0000 UTC m=+4084.452700834" lastFinishedPulling="2026-01-06 15:44:50.554774641 +0000 UTC m=+4087.182240969" observedRunningTime="2026-01-06 15:44:50.88173333 +0000 UTC m=+4087.509199648" watchObservedRunningTime="2026-01-06 15:44:50.896067392 +0000 UTC m=+4087.523533710" Jan 06 15:44:56 crc kubenswrapper[4744]: I0106 15:44:56.093382 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:56 crc kubenswrapper[4744]: I0106 15:44:56.094013 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:57 crc kubenswrapper[4744]: I0106 15:44:57.112233 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:57 crc kubenswrapper[4744]: I0106 15:44:57.189706 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:57 crc kubenswrapper[4744]: I0106 15:44:57.364873 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvf6j"] Jan 06 15:44:58 crc kubenswrapper[4744]: I0106 15:44:58.966303 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tvf6j" podUID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerName="registry-server" containerID="cri-o://8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83" gracePeriod=2 Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.581488 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.709517 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92csz\" (UniqueName: \"kubernetes.io/projected/a8e774c8-20c3-44db-87d3-9431bc0c6e13-kube-api-access-92csz\") pod \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.709739 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-utilities\") pod \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.709799 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-catalog-content\") pod \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\" (UID: \"a8e774c8-20c3-44db-87d3-9431bc0c6e13\") " Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.711386 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-utilities" (OuterVolumeSpecName: "utilities") pod "a8e774c8-20c3-44db-87d3-9431bc0c6e13" (UID: "a8e774c8-20c3-44db-87d3-9431bc0c6e13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.718427 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8e774c8-20c3-44db-87d3-9431bc0c6e13-kube-api-access-92csz" (OuterVolumeSpecName: "kube-api-access-92csz") pod "a8e774c8-20c3-44db-87d3-9431bc0c6e13" (UID: "a8e774c8-20c3-44db-87d3-9431bc0c6e13"). InnerVolumeSpecName "kube-api-access-92csz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.813590 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.813639 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92csz\" (UniqueName: \"kubernetes.io/projected/a8e774c8-20c3-44db-87d3-9431bc0c6e13-kube-api-access-92csz\") on node \"crc\" DevicePath \"\"" Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.970676 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a8e774c8-20c3-44db-87d3-9431bc0c6e13" (UID: "a8e774c8-20c3-44db-87d3-9431bc0c6e13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.986009 4744 generic.go:334] "Generic (PLEG): container finished" podID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerID="8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83" exitCode=0 Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.986068 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvf6j" event={"ID":"a8e774c8-20c3-44db-87d3-9431bc0c6e13","Type":"ContainerDied","Data":"8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83"} Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.986099 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tvf6j" Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.986189 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tvf6j" event={"ID":"a8e774c8-20c3-44db-87d3-9431bc0c6e13","Type":"ContainerDied","Data":"343b66815b3fa96efdbe955352db7c2a8ec83707179555d5a02ab113ee3c7ce7"} Jan 06 15:44:59 crc kubenswrapper[4744]: I0106 15:44:59.986228 4744 scope.go:117] "RemoveContainer" containerID="8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.024496 4744 scope.go:117] "RemoveContainer" containerID="02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.029045 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a8e774c8-20c3-44db-87d3-9431bc0c6e13-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.038105 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tvf6j"] Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.053452 4744 scope.go:117] "RemoveContainer" containerID="93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.060944 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tvf6j"] Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.123520 4744 scope.go:117] "RemoveContainer" containerID="8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83" Jan 06 15:45:00 crc kubenswrapper[4744]: E0106 15:45:00.124111 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83\": container with ID starting with 8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83 not found: ID does not exist" containerID="8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.124211 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83"} err="failed to get container status \"8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83\": rpc error: code = NotFound desc = could not find container \"8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83\": container with ID starting with 8afa1780a340a87c58eb34881300774242cba417dcd9a7ee3ff432164481ce83 not found: ID does not exist" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.124246 4744 scope.go:117] "RemoveContainer" containerID="02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69" Jan 06 15:45:00 crc kubenswrapper[4744]: E0106 15:45:00.124946 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69\": container with ID starting with 02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69 not found: ID does not exist" containerID="02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.124981 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69"} err="failed to get container status \"02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69\": rpc error: code = NotFound desc = could not find container \"02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69\": container with ID starting with 02ea7491a704b38271bccb56eada4933461def250205fffdeb72a2809124ab69 not found: ID does not exist" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.124999 4744 scope.go:117] "RemoveContainer" containerID="93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e" Jan 06 15:45:00 crc kubenswrapper[4744]: E0106 15:45:00.125287 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e\": container with ID starting with 93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e not found: ID does not exist" containerID="93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.125314 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e"} err="failed to get container status \"93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e\": rpc error: code = NotFound desc = could not find container \"93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e\": container with ID starting with 93a1ae4d3f614275cdf9f561cc2570eaef7a750e847f3534418cf4cec1d7645e not found: ID does not exist" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.172153 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct"] Jan 06 15:45:00 crc kubenswrapper[4744]: E0106 15:45:00.172855 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerName="registry-server" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.172877 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerName="registry-server" Jan 06 15:45:00 crc kubenswrapper[4744]: E0106 15:45:00.172911 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerName="extract-content" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.172919 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerName="extract-content" Jan 06 15:45:00 crc kubenswrapper[4744]: E0106 15:45:00.172943 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerName="extract-utilities" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.172952 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerName="extract-utilities" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.173242 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" containerName="registry-server" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.174357 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.203072 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.214689 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.225200 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct"] Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.335445 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2rgz\" (UniqueName: \"kubernetes.io/projected/99512364-e502-4a80-9054-9af9f683744d-kube-api-access-c2rgz\") pod \"collect-profiles-29461905-vccct\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.335561 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99512364-e502-4a80-9054-9af9f683744d-secret-volume\") pod \"collect-profiles-29461905-vccct\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.335587 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99512364-e502-4a80-9054-9af9f683744d-config-volume\") pod \"collect-profiles-29461905-vccct\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.439206 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99512364-e502-4a80-9054-9af9f683744d-secret-volume\") pod \"collect-profiles-29461905-vccct\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.439682 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99512364-e502-4a80-9054-9af9f683744d-config-volume\") pod \"collect-profiles-29461905-vccct\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.439939 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2rgz\" (UniqueName: \"kubernetes.io/projected/99512364-e502-4a80-9054-9af9f683744d-kube-api-access-c2rgz\") pod \"collect-profiles-29461905-vccct\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.440490 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99512364-e502-4a80-9054-9af9f683744d-config-volume\") pod \"collect-profiles-29461905-vccct\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.445294 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99512364-e502-4a80-9054-9af9f683744d-secret-volume\") pod \"collect-profiles-29461905-vccct\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.459606 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2rgz\" (UniqueName: \"kubernetes.io/projected/99512364-e502-4a80-9054-9af9f683744d-kube-api-access-c2rgz\") pod \"collect-profiles-29461905-vccct\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:00 crc kubenswrapper[4744]: I0106 15:45:00.517241 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:01 crc kubenswrapper[4744]: I0106 15:45:01.014079 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct"] Jan 06 15:45:01 crc kubenswrapper[4744]: I0106 15:45:01.729024 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8e774c8-20c3-44db-87d3-9431bc0c6e13" path="/var/lib/kubelet/pods/a8e774c8-20c3-44db-87d3-9431bc0c6e13/volumes" Jan 06 15:45:02 crc kubenswrapper[4744]: I0106 15:45:02.013696 4744 generic.go:334] "Generic (PLEG): container finished" podID="99512364-e502-4a80-9054-9af9f683744d" containerID="a9c69895f9613e32cac822451bb0e1741b2b7860db0511e0feab9e7714babeb5" exitCode=0 Jan 06 15:45:02 crc kubenswrapper[4744]: I0106 15:45:02.013773 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" event={"ID":"99512364-e502-4a80-9054-9af9f683744d","Type":"ContainerDied","Data":"a9c69895f9613e32cac822451bb0e1741b2b7860db0511e0feab9e7714babeb5"} Jan 06 15:45:02 crc kubenswrapper[4744]: I0106 15:45:02.014204 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" event={"ID":"99512364-e502-4a80-9054-9af9f683744d","Type":"ContainerStarted","Data":"5fe8ccaefc24d71b2f52d9fe40e032d7126f6f1909dea3e5a22ba4bbb37bb2ba"} Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.495944 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.617757 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2rgz\" (UniqueName: \"kubernetes.io/projected/99512364-e502-4a80-9054-9af9f683744d-kube-api-access-c2rgz\") pod \"99512364-e502-4a80-9054-9af9f683744d\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.618050 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99512364-e502-4a80-9054-9af9f683744d-secret-volume\") pod \"99512364-e502-4a80-9054-9af9f683744d\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.618110 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99512364-e502-4a80-9054-9af9f683744d-config-volume\") pod \"99512364-e502-4a80-9054-9af9f683744d\" (UID: \"99512364-e502-4a80-9054-9af9f683744d\") " Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.618843 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99512364-e502-4a80-9054-9af9f683744d-config-volume" (OuterVolumeSpecName: "config-volume") pod "99512364-e502-4a80-9054-9af9f683744d" (UID: "99512364-e502-4a80-9054-9af9f683744d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.619474 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99512364-e502-4a80-9054-9af9f683744d-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.632616 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99512364-e502-4a80-9054-9af9f683744d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "99512364-e502-4a80-9054-9af9f683744d" (UID: "99512364-e502-4a80-9054-9af9f683744d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.632874 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99512364-e502-4a80-9054-9af9f683744d-kube-api-access-c2rgz" (OuterVolumeSpecName: "kube-api-access-c2rgz") pod "99512364-e502-4a80-9054-9af9f683744d" (UID: "99512364-e502-4a80-9054-9af9f683744d"). InnerVolumeSpecName "kube-api-access-c2rgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.722954 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2rgz\" (UniqueName: \"kubernetes.io/projected/99512364-e502-4a80-9054-9af9f683744d-kube-api-access-c2rgz\") on node \"crc\" DevicePath \"\"" Jan 06 15:45:03 crc kubenswrapper[4744]: I0106 15:45:03.723245 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99512364-e502-4a80-9054-9af9f683744d-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 06 15:45:04 crc kubenswrapper[4744]: I0106 15:45:04.040627 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" event={"ID":"99512364-e502-4a80-9054-9af9f683744d","Type":"ContainerDied","Data":"5fe8ccaefc24d71b2f52d9fe40e032d7126f6f1909dea3e5a22ba4bbb37bb2ba"} Jan 06 15:45:04 crc kubenswrapper[4744]: I0106 15:45:04.040682 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fe8ccaefc24d71b2f52d9fe40e032d7126f6f1909dea3e5a22ba4bbb37bb2ba" Jan 06 15:45:04 crc kubenswrapper[4744]: I0106 15:45:04.040694 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct" Jan 06 15:45:04 crc kubenswrapper[4744]: I0106 15:45:04.603433 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2"] Jan 06 15:45:04 crc kubenswrapper[4744]: I0106 15:45:04.616133 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461860-rshx2"] Jan 06 15:45:05 crc kubenswrapper[4744]: I0106 15:45:05.736676 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa2592b1-dd21-4676-bea3-d032eb197ebb" path="/var/lib/kubelet/pods/aa2592b1-dd21-4676-bea3-d032eb197ebb/volumes" Jan 06 15:45:13 crc kubenswrapper[4744]: I0106 15:45:13.172001 4744 scope.go:117] "RemoveContainer" containerID="f6de9cf249d4b29fe4a377e8df6f3754b2ded2846bab01bb5deb176ad54c09a1" Jan 06 15:45:14 crc kubenswrapper[4744]: I0106 15:45:14.423594 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:45:14 crc kubenswrapper[4744]: I0106 15:45:14.423691 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:45:14 crc kubenswrapper[4744]: I0106 15:45:14.423789 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:45:14 crc kubenswrapper[4744]: I0106 15:45:14.425437 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc3d927402f59963ca59e24827db41e1cf066041e70d1875be675145db361865"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:45:14 crc kubenswrapper[4744]: I0106 15:45:14.425851 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://bc3d927402f59963ca59e24827db41e1cf066041e70d1875be675145db361865" gracePeriod=600 Jan 06 15:45:15 crc kubenswrapper[4744]: I0106 15:45:15.215341 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="bc3d927402f59963ca59e24827db41e1cf066041e70d1875be675145db361865" exitCode=0 Jan 06 15:45:15 crc kubenswrapper[4744]: I0106 15:45:15.215425 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"bc3d927402f59963ca59e24827db41e1cf066041e70d1875be675145db361865"} Jan 06 15:45:15 crc kubenswrapper[4744]: I0106 15:45:15.215728 4744 scope.go:117] "RemoveContainer" containerID="e6c13f703ae01d164d7c952ae8e5c3a7840cd5d043af68ed3330664fc9741a1d" Jan 06 15:45:16 crc kubenswrapper[4744]: I0106 15:45:16.234418 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a"} Jan 06 15:47:44 crc kubenswrapper[4744]: I0106 15:47:44.423800 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:47:44 crc kubenswrapper[4744]: I0106 15:47:44.424284 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:48:14 crc kubenswrapper[4744]: I0106 15:48:14.423513 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:48:14 crc kubenswrapper[4744]: I0106 15:48:14.424156 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:48:44 crc kubenswrapper[4744]: I0106 15:48:44.423923 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:48:44 crc kubenswrapper[4744]: I0106 15:48:44.424474 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:48:44 crc kubenswrapper[4744]: I0106 15:48:44.424535 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:48:44 crc kubenswrapper[4744]: I0106 15:48:44.425937 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:48:44 crc kubenswrapper[4744]: I0106 15:48:44.426043 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" gracePeriod=600 Jan 06 15:48:44 crc kubenswrapper[4744]: E0106 15:48:44.559030 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:48:45 crc kubenswrapper[4744]: I0106 15:48:45.083734 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" exitCode=0 Jan 06 15:48:45 crc kubenswrapper[4744]: I0106 15:48:45.083811 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a"} Jan 06 15:48:45 crc kubenswrapper[4744]: I0106 15:48:45.083876 4744 scope.go:117] "RemoveContainer" containerID="bc3d927402f59963ca59e24827db41e1cf066041e70d1875be675145db361865" Jan 06 15:48:45 crc kubenswrapper[4744]: I0106 15:48:45.085342 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:48:45 crc kubenswrapper[4744]: E0106 15:48:45.086374 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:48:56 crc kubenswrapper[4744]: I0106 15:48:56.713395 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:48:56 crc kubenswrapper[4744]: E0106 15:48:56.714112 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:49:07 crc kubenswrapper[4744]: I0106 15:49:07.711663 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:49:07 crc kubenswrapper[4744]: E0106 15:49:07.712468 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:49:18 crc kubenswrapper[4744]: I0106 15:49:18.711601 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:49:18 crc kubenswrapper[4744]: E0106 15:49:18.713029 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:49:33 crc kubenswrapper[4744]: I0106 15:49:33.725608 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:49:33 crc kubenswrapper[4744]: E0106 15:49:33.726739 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:49:47 crc kubenswrapper[4744]: I0106 15:49:47.712508 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:49:47 crc kubenswrapper[4744]: E0106 15:49:47.713693 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:49:59 crc kubenswrapper[4744]: I0106 15:49:59.712459 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:49:59 crc kubenswrapper[4744]: E0106 15:49:59.713616 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:50:11 crc kubenswrapper[4744]: I0106 15:50:11.711525 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:50:11 crc kubenswrapper[4744]: E0106 15:50:11.712498 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:50:22 crc kubenswrapper[4744]: I0106 15:50:22.712584 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:50:22 crc kubenswrapper[4744]: E0106 15:50:22.713785 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:50:37 crc kubenswrapper[4744]: I0106 15:50:37.711632 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:50:37 crc kubenswrapper[4744]: E0106 15:50:37.712700 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:50:52 crc kubenswrapper[4744]: I0106 15:50:52.711116 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:50:52 crc kubenswrapper[4744]: E0106 15:50:52.713671 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:51:05 crc kubenswrapper[4744]: I0106 15:51:05.712102 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:51:05 crc kubenswrapper[4744]: E0106 15:51:05.712999 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:51:20 crc kubenswrapper[4744]: I0106 15:51:20.711184 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:51:20 crc kubenswrapper[4744]: E0106 15:51:20.712027 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:51:31 crc kubenswrapper[4744]: I0106 15:51:31.711176 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:51:31 crc kubenswrapper[4744]: E0106 15:51:31.711828 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:51:46 crc kubenswrapper[4744]: I0106 15:51:46.714083 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:51:46 crc kubenswrapper[4744]: E0106 15:51:46.714992 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.190483 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c4tbk"] Jan 06 15:51:55 crc kubenswrapper[4744]: E0106 15:51:55.191459 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99512364-e502-4a80-9054-9af9f683744d" containerName="collect-profiles" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.191471 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="99512364-e502-4a80-9054-9af9f683744d" containerName="collect-profiles" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.191711 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="99512364-e502-4a80-9054-9af9f683744d" containerName="collect-profiles" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.193341 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.203699 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4tbk"] Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.355727 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxrn5\" (UniqueName: \"kubernetes.io/projected/c05bf95b-0559-4acb-8908-a0063bf51ec2-kube-api-access-vxrn5\") pod \"certified-operators-c4tbk\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.356023 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-utilities\") pod \"certified-operators-c4tbk\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.356088 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-catalog-content\") pod \"certified-operators-c4tbk\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.458369 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxrn5\" (UniqueName: \"kubernetes.io/projected/c05bf95b-0559-4acb-8908-a0063bf51ec2-kube-api-access-vxrn5\") pod \"certified-operators-c4tbk\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.458420 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-utilities\") pod \"certified-operators-c4tbk\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.458488 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-catalog-content\") pod \"certified-operators-c4tbk\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.458909 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-utilities\") pod \"certified-operators-c4tbk\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.458940 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-catalog-content\") pod \"certified-operators-c4tbk\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.485836 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxrn5\" (UniqueName: \"kubernetes.io/projected/c05bf95b-0559-4acb-8908-a0063bf51ec2-kube-api-access-vxrn5\") pod \"certified-operators-c4tbk\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:55 crc kubenswrapper[4744]: I0106 15:51:55.512089 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:51:56 crc kubenswrapper[4744]: I0106 15:51:56.022754 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c4tbk"] Jan 06 15:51:56 crc kubenswrapper[4744]: I0106 15:51:56.726940 4744 generic.go:334] "Generic (PLEG): container finished" podID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerID="7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31" exitCode=0 Jan 06 15:51:56 crc kubenswrapper[4744]: I0106 15:51:56.727321 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4tbk" event={"ID":"c05bf95b-0559-4acb-8908-a0063bf51ec2","Type":"ContainerDied","Data":"7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31"} Jan 06 15:51:56 crc kubenswrapper[4744]: I0106 15:51:56.727359 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4tbk" event={"ID":"c05bf95b-0559-4acb-8908-a0063bf51ec2","Type":"ContainerStarted","Data":"d8cbd4783d7cb543591c0265a0193ae52919da7042704b81192d5311e351e8fa"} Jan 06 15:51:56 crc kubenswrapper[4744]: I0106 15:51:56.730406 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 15:51:57 crc kubenswrapper[4744]: I0106 15:51:57.755804 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4tbk" event={"ID":"c05bf95b-0559-4acb-8908-a0063bf51ec2","Type":"ContainerStarted","Data":"8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a"} Jan 06 15:51:59 crc kubenswrapper[4744]: I0106 15:51:59.782432 4744 generic.go:334] "Generic (PLEG): container finished" podID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerID="8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a" exitCode=0 Jan 06 15:51:59 crc kubenswrapper[4744]: I0106 15:51:59.782550 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4tbk" event={"ID":"c05bf95b-0559-4acb-8908-a0063bf51ec2","Type":"ContainerDied","Data":"8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a"} Jan 06 15:52:00 crc kubenswrapper[4744]: I0106 15:52:00.795762 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4tbk" event={"ID":"c05bf95b-0559-4acb-8908-a0063bf51ec2","Type":"ContainerStarted","Data":"ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549"} Jan 06 15:52:00 crc kubenswrapper[4744]: I0106 15:52:00.823724 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c4tbk" podStartSLOduration=2.358648401 podStartE2EDuration="5.823692665s" podCreationTimestamp="2026-01-06 15:51:55 +0000 UTC" firstStartedPulling="2026-01-06 15:51:56.729596207 +0000 UTC m=+4513.357062565" lastFinishedPulling="2026-01-06 15:52:00.194640511 +0000 UTC m=+4516.822106829" observedRunningTime="2026-01-06 15:52:00.816639676 +0000 UTC m=+4517.444106004" watchObservedRunningTime="2026-01-06 15:52:00.823692665 +0000 UTC m=+4517.451159023" Jan 06 15:52:01 crc kubenswrapper[4744]: I0106 15:52:01.713503 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:52:01 crc kubenswrapper[4744]: E0106 15:52:01.715454 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.382938 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mldhz"] Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.387232 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.404302 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mldhz"] Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.496487 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-catalog-content\") pod \"redhat-marketplace-mldhz\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.497149 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-utilities\") pod \"redhat-marketplace-mldhz\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.497413 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2lz8\" (UniqueName: \"kubernetes.io/projected/a2fb0a32-158f-4dfc-9114-67d303c14844-kube-api-access-d2lz8\") pod \"redhat-marketplace-mldhz\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.600112 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-utilities\") pod \"redhat-marketplace-mldhz\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.600186 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2lz8\" (UniqueName: \"kubernetes.io/projected/a2fb0a32-158f-4dfc-9114-67d303c14844-kube-api-access-d2lz8\") pod \"redhat-marketplace-mldhz\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.600275 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-catalog-content\") pod \"redhat-marketplace-mldhz\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.600904 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-utilities\") pod \"redhat-marketplace-mldhz\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.601104 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-catalog-content\") pod \"redhat-marketplace-mldhz\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.620730 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2lz8\" (UniqueName: \"kubernetes.io/projected/a2fb0a32-158f-4dfc-9114-67d303c14844-kube-api-access-d2lz8\") pod \"redhat-marketplace-mldhz\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:04 crc kubenswrapper[4744]: I0106 15:52:04.711238 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:05 crc kubenswrapper[4744]: I0106 15:52:05.295794 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mldhz"] Jan 06 15:52:05 crc kubenswrapper[4744]: W0106 15:52:05.299754 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2fb0a32_158f_4dfc_9114_67d303c14844.slice/crio-537ca832558d6b1bd1859d88eed633754d5767a4858da97765e4036854904ddc WatchSource:0}: Error finding container 537ca832558d6b1bd1859d88eed633754d5767a4858da97765e4036854904ddc: Status 404 returned error can't find the container with id 537ca832558d6b1bd1859d88eed633754d5767a4858da97765e4036854904ddc Jan 06 15:52:05 crc kubenswrapper[4744]: I0106 15:52:05.513014 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:52:05 crc kubenswrapper[4744]: I0106 15:52:05.513521 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:52:05 crc kubenswrapper[4744]: I0106 15:52:05.580862 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:52:05 crc kubenswrapper[4744]: I0106 15:52:05.863771 4744 generic.go:334] "Generic (PLEG): container finished" podID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerID="cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38" exitCode=0 Jan 06 15:52:05 crc kubenswrapper[4744]: I0106 15:52:05.863848 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mldhz" event={"ID":"a2fb0a32-158f-4dfc-9114-67d303c14844","Type":"ContainerDied","Data":"cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38"} Jan 06 15:52:05 crc kubenswrapper[4744]: I0106 15:52:05.863886 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mldhz" event={"ID":"a2fb0a32-158f-4dfc-9114-67d303c14844","Type":"ContainerStarted","Data":"537ca832558d6b1bd1859d88eed633754d5767a4858da97765e4036854904ddc"} Jan 06 15:52:05 crc kubenswrapper[4744]: I0106 15:52:05.930874 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:52:07 crc kubenswrapper[4744]: I0106 15:52:07.891848 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mldhz" event={"ID":"a2fb0a32-158f-4dfc-9114-67d303c14844","Type":"ContainerStarted","Data":"2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992"} Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.178271 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4tbk"] Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.178520 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c4tbk" podUID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerName="registry-server" containerID="cri-o://ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549" gracePeriod=2 Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.732559 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.905558 4744 generic.go:334] "Generic (PLEG): container finished" podID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerID="2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992" exitCode=0 Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.905604 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mldhz" event={"ID":"a2fb0a32-158f-4dfc-9114-67d303c14844","Type":"ContainerDied","Data":"2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992"} Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.907571 4744 generic.go:334] "Generic (PLEG): container finished" podID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerID="ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549" exitCode=0 Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.907626 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4tbk" event={"ID":"c05bf95b-0559-4acb-8908-a0063bf51ec2","Type":"ContainerDied","Data":"ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549"} Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.907632 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c4tbk" Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.907657 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c4tbk" event={"ID":"c05bf95b-0559-4acb-8908-a0063bf51ec2","Type":"ContainerDied","Data":"d8cbd4783d7cb543591c0265a0193ae52919da7042704b81192d5311e351e8fa"} Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.907679 4744 scope.go:117] "RemoveContainer" containerID="ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549" Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.918085 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-utilities\") pod \"c05bf95b-0559-4acb-8908-a0063bf51ec2\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.918298 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxrn5\" (UniqueName: \"kubernetes.io/projected/c05bf95b-0559-4acb-8908-a0063bf51ec2-kube-api-access-vxrn5\") pod \"c05bf95b-0559-4acb-8908-a0063bf51ec2\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.918343 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-catalog-content\") pod \"c05bf95b-0559-4acb-8908-a0063bf51ec2\" (UID: \"c05bf95b-0559-4acb-8908-a0063bf51ec2\") " Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.919274 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-utilities" (OuterVolumeSpecName: "utilities") pod "c05bf95b-0559-4acb-8908-a0063bf51ec2" (UID: "c05bf95b-0559-4acb-8908-a0063bf51ec2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.931529 4744 scope.go:117] "RemoveContainer" containerID="8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a" Jan 06 15:52:08 crc kubenswrapper[4744]: I0106 15:52:08.990875 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c05bf95b-0559-4acb-8908-a0063bf51ec2" (UID: "c05bf95b-0559-4acb-8908-a0063bf51ec2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.022480 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.022524 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c05bf95b-0559-4acb-8908-a0063bf51ec2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.356217 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c05bf95b-0559-4acb-8908-a0063bf51ec2-kube-api-access-vxrn5" (OuterVolumeSpecName: "kube-api-access-vxrn5") pod "c05bf95b-0559-4acb-8908-a0063bf51ec2" (UID: "c05bf95b-0559-4acb-8908-a0063bf51ec2"). InnerVolumeSpecName "kube-api-access-vxrn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.381130 4744 scope.go:117] "RemoveContainer" containerID="7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.431406 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxrn5\" (UniqueName: \"kubernetes.io/projected/c05bf95b-0559-4acb-8908-a0063bf51ec2-kube-api-access-vxrn5\") on node \"crc\" DevicePath \"\"" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.574941 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c4tbk"] Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.587416 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c4tbk"] Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.683015 4744 scope.go:117] "RemoveContainer" containerID="ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549" Jan 06 15:52:09 crc kubenswrapper[4744]: E0106 15:52:09.683583 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549\": container with ID starting with ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549 not found: ID does not exist" containerID="ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.683644 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549"} err="failed to get container status \"ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549\": rpc error: code = NotFound desc = could not find container \"ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549\": container with ID starting with ce3da07b6d84472a1bad1e8143e98d2fd5873269797896c9b0f11cb605635549 not found: ID does not exist" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.683688 4744 scope.go:117] "RemoveContainer" containerID="8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a" Jan 06 15:52:09 crc kubenswrapper[4744]: E0106 15:52:09.684331 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a\": container with ID starting with 8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a not found: ID does not exist" containerID="8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.684360 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a"} err="failed to get container status \"8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a\": rpc error: code = NotFound desc = could not find container \"8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a\": container with ID starting with 8150569306e9d9ace3c5ba7cf7dbaa2e3aadb635aa60b1adebe5afa807cb464a not found: ID does not exist" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.684382 4744 scope.go:117] "RemoveContainer" containerID="7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31" Jan 06 15:52:09 crc kubenswrapper[4744]: E0106 15:52:09.684723 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31\": container with ID starting with 7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31 not found: ID does not exist" containerID="7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.684740 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31"} err="failed to get container status \"7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31\": rpc error: code = NotFound desc = could not find container \"7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31\": container with ID starting with 7ba08916f3b0899e28add51077c752d4e60e0af746fb7eff9db7e97401ecae31 not found: ID does not exist" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.730211 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c05bf95b-0559-4acb-8908-a0063bf51ec2" path="/var/lib/kubelet/pods/c05bf95b-0559-4acb-8908-a0063bf51ec2/volumes" Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.920720 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mldhz" event={"ID":"a2fb0a32-158f-4dfc-9114-67d303c14844","Type":"ContainerStarted","Data":"f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1"} Jan 06 15:52:09 crc kubenswrapper[4744]: I0106 15:52:09.939587 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mldhz" podStartSLOduration=2.3813569279999998 podStartE2EDuration="5.939570716s" podCreationTimestamp="2026-01-06 15:52:04 +0000 UTC" firstStartedPulling="2026-01-06 15:52:05.866356056 +0000 UTC m=+4522.493822374" lastFinishedPulling="2026-01-06 15:52:09.424569844 +0000 UTC m=+4526.052036162" observedRunningTime="2026-01-06 15:52:09.935830436 +0000 UTC m=+4526.563296754" watchObservedRunningTime="2026-01-06 15:52:09.939570716 +0000 UTC m=+4526.567037034" Jan 06 15:52:14 crc kubenswrapper[4744]: I0106 15:52:14.711544 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:14 crc kubenswrapper[4744]: I0106 15:52:14.712599 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:14 crc kubenswrapper[4744]: I0106 15:52:14.777726 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:15 crc kubenswrapper[4744]: I0106 15:52:15.357994 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:16 crc kubenswrapper[4744]: I0106 15:52:16.711826 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:52:16 crc kubenswrapper[4744]: E0106 15:52:16.713297 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:52:18 crc kubenswrapper[4744]: I0106 15:52:18.360601 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mldhz"] Jan 06 15:52:18 crc kubenswrapper[4744]: I0106 15:52:18.361587 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mldhz" podUID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerName="registry-server" containerID="cri-o://f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1" gracePeriod=2 Jan 06 15:52:18 crc kubenswrapper[4744]: E0106 15:52:18.613704 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2fb0a32_158f_4dfc_9114_67d303c14844.slice/crio-f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2fb0a32_158f_4dfc_9114_67d303c14844.slice/crio-conmon-f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1.scope\": RecentStats: unable to find data in memory cache]" Jan 06 15:52:18 crc kubenswrapper[4744]: I0106 15:52:18.935246 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.082373 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2lz8\" (UniqueName: \"kubernetes.io/projected/a2fb0a32-158f-4dfc-9114-67d303c14844-kube-api-access-d2lz8\") pod \"a2fb0a32-158f-4dfc-9114-67d303c14844\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.082493 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-catalog-content\") pod \"a2fb0a32-158f-4dfc-9114-67d303c14844\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.082579 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-utilities\") pod \"a2fb0a32-158f-4dfc-9114-67d303c14844\" (UID: \"a2fb0a32-158f-4dfc-9114-67d303c14844\") " Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.083307 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-utilities" (OuterVolumeSpecName: "utilities") pod "a2fb0a32-158f-4dfc-9114-67d303c14844" (UID: "a2fb0a32-158f-4dfc-9114-67d303c14844"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.090015 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2fb0a32-158f-4dfc-9114-67d303c14844-kube-api-access-d2lz8" (OuterVolumeSpecName: "kube-api-access-d2lz8") pod "a2fb0a32-158f-4dfc-9114-67d303c14844" (UID: "a2fb0a32-158f-4dfc-9114-67d303c14844"). InnerVolumeSpecName "kube-api-access-d2lz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.124121 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2fb0a32-158f-4dfc-9114-67d303c14844" (UID: "a2fb0a32-158f-4dfc-9114-67d303c14844"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.185397 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2lz8\" (UniqueName: \"kubernetes.io/projected/a2fb0a32-158f-4dfc-9114-67d303c14844-kube-api-access-d2lz8\") on node \"crc\" DevicePath \"\"" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.185436 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.185445 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2fb0a32-158f-4dfc-9114-67d303c14844-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.350445 4744 generic.go:334] "Generic (PLEG): container finished" podID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerID="f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1" exitCode=0 Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.350509 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mldhz" event={"ID":"a2fb0a32-158f-4dfc-9114-67d303c14844","Type":"ContainerDied","Data":"f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1"} Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.350537 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mldhz" event={"ID":"a2fb0a32-158f-4dfc-9114-67d303c14844","Type":"ContainerDied","Data":"537ca832558d6b1bd1859d88eed633754d5767a4858da97765e4036854904ddc"} Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.350555 4744 scope.go:117] "RemoveContainer" containerID="f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.350562 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mldhz" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.404857 4744 scope.go:117] "RemoveContainer" containerID="2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.405730 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mldhz"] Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.416353 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mldhz"] Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.436925 4744 scope.go:117] "RemoveContainer" containerID="cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.512905 4744 scope.go:117] "RemoveContainer" containerID="f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1" Jan 06 15:52:19 crc kubenswrapper[4744]: E0106 15:52:19.513623 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1\": container with ID starting with f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1 not found: ID does not exist" containerID="f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.513666 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1"} err="failed to get container status \"f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1\": rpc error: code = NotFound desc = could not find container \"f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1\": container with ID starting with f13d3f7ffbe2b4c06c6b76070b42ad2bf8cc0a62a358da2c4e6dbdc7624ad7f1 not found: ID does not exist" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.513693 4744 scope.go:117] "RemoveContainer" containerID="2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992" Jan 06 15:52:19 crc kubenswrapper[4744]: E0106 15:52:19.514192 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992\": container with ID starting with 2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992 not found: ID does not exist" containerID="2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.514222 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992"} err="failed to get container status \"2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992\": rpc error: code = NotFound desc = could not find container \"2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992\": container with ID starting with 2d4e2b958483bdc0c7b3f7b6d0bde5254abef4250644156783dcbf4d3c379992 not found: ID does not exist" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.514243 4744 scope.go:117] "RemoveContainer" containerID="cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38" Jan 06 15:52:19 crc kubenswrapper[4744]: E0106 15:52:19.514650 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38\": container with ID starting with cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38 not found: ID does not exist" containerID="cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.514694 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38"} err="failed to get container status \"cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38\": rpc error: code = NotFound desc = could not find container \"cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38\": container with ID starting with cfff92486109c563c4b87653c75daa01e68994cd616fd0b10f192aa91c4efc38 not found: ID does not exist" Jan 06 15:52:19 crc kubenswrapper[4744]: I0106 15:52:19.724131 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2fb0a32-158f-4dfc-9114-67d303c14844" path="/var/lib/kubelet/pods/a2fb0a32-158f-4dfc-9114-67d303c14844/volumes" Jan 06 15:52:27 crc kubenswrapper[4744]: I0106 15:52:27.711977 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:52:27 crc kubenswrapper[4744]: E0106 15:52:27.713282 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:52:42 crc kubenswrapper[4744]: I0106 15:52:42.713028 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:52:42 crc kubenswrapper[4744]: E0106 15:52:42.714379 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:52:57 crc kubenswrapper[4744]: I0106 15:52:57.712384 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:52:57 crc kubenswrapper[4744]: E0106 15:52:57.714045 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:53:09 crc kubenswrapper[4744]: I0106 15:53:09.711389 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:53:09 crc kubenswrapper[4744]: E0106 15:53:09.712549 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:53:15 crc kubenswrapper[4744]: I0106 15:53:15.049989 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-mmvx7" podUID="4d920867-2ba6-42f6-acc3-266cbfbc7716" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 15:53:24 crc kubenswrapper[4744]: E0106 15:53:24.312316 4744 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.22:51450->38.102.83.22:45165: read tcp 38.102.83.22:51450->38.102.83.22:45165: read: connection reset by peer Jan 06 15:53:24 crc kubenswrapper[4744]: E0106 15:53:24.313615 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.22:51450->38.102.83.22:45165: write tcp 38.102.83.22:51450->38.102.83.22:45165: write: broken pipe Jan 06 15:53:24 crc kubenswrapper[4744]: I0106 15:53:24.711995 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:53:24 crc kubenswrapper[4744]: E0106 15:53:24.712644 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.790596 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f9kdq"] Jan 06 15:53:32 crc kubenswrapper[4744]: E0106 15:53:32.791912 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerName="registry-server" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.791934 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerName="registry-server" Jan 06 15:53:32 crc kubenswrapper[4744]: E0106 15:53:32.791965 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerName="extract-utilities" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.791978 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerName="extract-utilities" Jan 06 15:53:32 crc kubenswrapper[4744]: E0106 15:53:32.792003 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerName="extract-content" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.792018 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerName="extract-content" Jan 06 15:53:32 crc kubenswrapper[4744]: E0106 15:53:32.792044 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerName="extract-utilities" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.792056 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerName="extract-utilities" Jan 06 15:53:32 crc kubenswrapper[4744]: E0106 15:53:32.792076 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerName="registry-server" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.792087 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerName="registry-server" Jan 06 15:53:32 crc kubenswrapper[4744]: E0106 15:53:32.792117 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerName="extract-content" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.792128 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerName="extract-content" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.792617 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2fb0a32-158f-4dfc-9114-67d303c14844" containerName="registry-server" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.792672 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="c05bf95b-0559-4acb-8908-a0063bf51ec2" containerName="registry-server" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.795755 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.807929 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9kdq"] Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.906920 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7zq6\" (UniqueName: \"kubernetes.io/projected/8fe2d148-20af-45c5-ad68-5f2508662eca-kube-api-access-m7zq6\") pod \"redhat-operators-f9kdq\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.907037 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-utilities\") pod \"redhat-operators-f9kdq\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:32 crc kubenswrapper[4744]: I0106 15:53:32.907174 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-catalog-content\") pod \"redhat-operators-f9kdq\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:33 crc kubenswrapper[4744]: I0106 15:53:33.009064 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-catalog-content\") pod \"redhat-operators-f9kdq\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:33 crc kubenswrapper[4744]: I0106 15:53:33.009185 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7zq6\" (UniqueName: \"kubernetes.io/projected/8fe2d148-20af-45c5-ad68-5f2508662eca-kube-api-access-m7zq6\") pod \"redhat-operators-f9kdq\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:33 crc kubenswrapper[4744]: I0106 15:53:33.009258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-utilities\") pod \"redhat-operators-f9kdq\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:33 crc kubenswrapper[4744]: I0106 15:53:33.009745 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-catalog-content\") pod \"redhat-operators-f9kdq\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:33 crc kubenswrapper[4744]: I0106 15:53:33.009761 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-utilities\") pod \"redhat-operators-f9kdq\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:33 crc kubenswrapper[4744]: I0106 15:53:33.030039 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7zq6\" (UniqueName: \"kubernetes.io/projected/8fe2d148-20af-45c5-ad68-5f2508662eca-kube-api-access-m7zq6\") pod \"redhat-operators-f9kdq\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:33 crc kubenswrapper[4744]: I0106 15:53:33.144360 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:33 crc kubenswrapper[4744]: I0106 15:53:33.683254 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f9kdq"] Jan 06 15:53:34 crc kubenswrapper[4744]: I0106 15:53:34.333615 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9kdq" event={"ID":"8fe2d148-20af-45c5-ad68-5f2508662eca","Type":"ContainerStarted","Data":"23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346"} Jan 06 15:53:34 crc kubenswrapper[4744]: I0106 15:53:34.334223 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9kdq" event={"ID":"8fe2d148-20af-45c5-ad68-5f2508662eca","Type":"ContainerStarted","Data":"de7e7c9f9742289c84c7d9f1fe6329fa41805e76fd1a9d0725d40e3d5513991d"} Jan 06 15:53:35 crc kubenswrapper[4744]: I0106 15:53:35.344490 4744 generic.go:334] "Generic (PLEG): container finished" podID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerID="23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346" exitCode=0 Jan 06 15:53:35 crc kubenswrapper[4744]: I0106 15:53:35.344609 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9kdq" event={"ID":"8fe2d148-20af-45c5-ad68-5f2508662eca","Type":"ContainerDied","Data":"23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346"} Jan 06 15:53:36 crc kubenswrapper[4744]: I0106 15:53:36.360351 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9kdq" event={"ID":"8fe2d148-20af-45c5-ad68-5f2508662eca","Type":"ContainerStarted","Data":"df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8"} Jan 06 15:53:39 crc kubenswrapper[4744]: I0106 15:53:39.405190 4744 generic.go:334] "Generic (PLEG): container finished" podID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerID="df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8" exitCode=0 Jan 06 15:53:39 crc kubenswrapper[4744]: I0106 15:53:39.405207 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9kdq" event={"ID":"8fe2d148-20af-45c5-ad68-5f2508662eca","Type":"ContainerDied","Data":"df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8"} Jan 06 15:53:39 crc kubenswrapper[4744]: I0106 15:53:39.711554 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:53:39 crc kubenswrapper[4744]: E0106 15:53:39.711943 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 15:53:40 crc kubenswrapper[4744]: I0106 15:53:40.420622 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9kdq" event={"ID":"8fe2d148-20af-45c5-ad68-5f2508662eca","Type":"ContainerStarted","Data":"e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242"} Jan 06 15:53:40 crc kubenswrapper[4744]: I0106 15:53:40.457703 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f9kdq" podStartSLOduration=2.935167648 podStartE2EDuration="8.457678931s" podCreationTimestamp="2026-01-06 15:53:32 +0000 UTC" firstStartedPulling="2026-01-06 15:53:34.336073737 +0000 UTC m=+4610.963540055" lastFinishedPulling="2026-01-06 15:53:39.85858502 +0000 UTC m=+4616.486051338" observedRunningTime="2026-01-06 15:53:40.44230716 +0000 UTC m=+4617.069773518" watchObservedRunningTime="2026-01-06 15:53:40.457678931 +0000 UTC m=+4617.085145259" Jan 06 15:53:43 crc kubenswrapper[4744]: I0106 15:53:43.144766 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:43 crc kubenswrapper[4744]: I0106 15:53:43.145665 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:44 crc kubenswrapper[4744]: I0106 15:53:44.230061 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f9kdq" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerName="registry-server" probeResult="failure" output=< Jan 06 15:53:44 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 15:53:44 crc kubenswrapper[4744]: > Jan 06 15:53:53 crc kubenswrapper[4744]: I0106 15:53:53.226984 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:53 crc kubenswrapper[4744]: I0106 15:53:53.294054 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:53 crc kubenswrapper[4744]: I0106 15:53:53.475659 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9kdq"] Jan 06 15:53:54 crc kubenswrapper[4744]: I0106 15:53:54.613503 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f9kdq" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerName="registry-server" containerID="cri-o://e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242" gracePeriod=2 Jan 06 15:53:54 crc kubenswrapper[4744]: I0106 15:53:54.712202 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.105784 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.208253 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-utilities\") pod \"8fe2d148-20af-45c5-ad68-5f2508662eca\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.208618 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7zq6\" (UniqueName: \"kubernetes.io/projected/8fe2d148-20af-45c5-ad68-5f2508662eca-kube-api-access-m7zq6\") pod \"8fe2d148-20af-45c5-ad68-5f2508662eca\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.208736 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-catalog-content\") pod \"8fe2d148-20af-45c5-ad68-5f2508662eca\" (UID: \"8fe2d148-20af-45c5-ad68-5f2508662eca\") " Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.210182 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-utilities" (OuterVolumeSpecName: "utilities") pod "8fe2d148-20af-45c5-ad68-5f2508662eca" (UID: "8fe2d148-20af-45c5-ad68-5f2508662eca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.217399 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe2d148-20af-45c5-ad68-5f2508662eca-kube-api-access-m7zq6" (OuterVolumeSpecName: "kube-api-access-m7zq6") pod "8fe2d148-20af-45c5-ad68-5f2508662eca" (UID: "8fe2d148-20af-45c5-ad68-5f2508662eca"). InnerVolumeSpecName "kube-api-access-m7zq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.322935 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.323196 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7zq6\" (UniqueName: \"kubernetes.io/projected/8fe2d148-20af-45c5-ad68-5f2508662eca-kube-api-access-m7zq6\") on node \"crc\" DevicePath \"\"" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.369345 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fe2d148-20af-45c5-ad68-5f2508662eca" (UID: "8fe2d148-20af-45c5-ad68-5f2508662eca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.424982 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe2d148-20af-45c5-ad68-5f2508662eca-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.637113 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"de1ccb05de866c492da3b9c727016ce800e411146c0daf1657febcb845d3ec8c"} Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.642007 4744 generic.go:334] "Generic (PLEG): container finished" podID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerID="e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242" exitCode=0 Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.642056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9kdq" event={"ID":"8fe2d148-20af-45c5-ad68-5f2508662eca","Type":"ContainerDied","Data":"e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242"} Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.642095 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f9kdq" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.642106 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f9kdq" event={"ID":"8fe2d148-20af-45c5-ad68-5f2508662eca","Type":"ContainerDied","Data":"de7e7c9f9742289c84c7d9f1fe6329fa41805e76fd1a9d0725d40e3d5513991d"} Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.642131 4744 scope.go:117] "RemoveContainer" containerID="e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.711959 4744 scope.go:117] "RemoveContainer" containerID="df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.730680 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f9kdq"] Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.730714 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f9kdq"] Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.773123 4744 scope.go:117] "RemoveContainer" containerID="23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.819426 4744 scope.go:117] "RemoveContainer" containerID="e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242" Jan 06 15:53:55 crc kubenswrapper[4744]: E0106 15:53:55.819871 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242\": container with ID starting with e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242 not found: ID does not exist" containerID="e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.819900 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242"} err="failed to get container status \"e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242\": rpc error: code = NotFound desc = could not find container \"e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242\": container with ID starting with e52cf05280dbdd6c637c67a11298aa83be2ef9b15dcd84f8db11e464086a9242 not found: ID does not exist" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.819938 4744 scope.go:117] "RemoveContainer" containerID="df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8" Jan 06 15:53:55 crc kubenswrapper[4744]: E0106 15:53:55.823803 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8\": container with ID starting with df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8 not found: ID does not exist" containerID="df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.823864 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8"} err="failed to get container status \"df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8\": rpc error: code = NotFound desc = could not find container \"df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8\": container with ID starting with df6768887b2ef596a88a91d22eaf8558c9bfc3a2f7673cf61da7cca3bcae0ae8 not found: ID does not exist" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.823930 4744 scope.go:117] "RemoveContainer" containerID="23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346" Jan 06 15:53:55 crc kubenswrapper[4744]: E0106 15:53:55.825012 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346\": container with ID starting with 23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346 not found: ID does not exist" containerID="23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346" Jan 06 15:53:55 crc kubenswrapper[4744]: I0106 15:53:55.825067 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346"} err="failed to get container status \"23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346\": rpc error: code = NotFound desc = could not find container \"23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346\": container with ID starting with 23a822af2c4e8f68d933e1b3d252d598ddcc66caca59f8dc324060761e123346 not found: ID does not exist" Jan 06 15:53:57 crc kubenswrapper[4744]: I0106 15:53:57.730039 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" path="/var/lib/kubelet/pods/8fe2d148-20af-45c5-ad68-5f2508662eca/volumes" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.802047 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5rwgj"] Jan 06 15:55:11 crc kubenswrapper[4744]: E0106 15:55:11.803403 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerName="extract-utilities" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.803424 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerName="extract-utilities" Jan 06 15:55:11 crc kubenswrapper[4744]: E0106 15:55:11.803457 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerName="extract-content" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.803470 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerName="extract-content" Jan 06 15:55:11 crc kubenswrapper[4744]: E0106 15:55:11.803523 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerName="registry-server" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.803535 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerName="registry-server" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.803960 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe2d148-20af-45c5-ad68-5f2508662eca" containerName="registry-server" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.806902 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.826086 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rwgj"] Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.868825 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-catalog-content\") pod \"community-operators-5rwgj\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.868891 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwp6b\" (UniqueName: \"kubernetes.io/projected/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-kube-api-access-xwp6b\") pod \"community-operators-5rwgj\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.868954 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-utilities\") pod \"community-operators-5rwgj\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.971113 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-catalog-content\") pod \"community-operators-5rwgj\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.971181 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwp6b\" (UniqueName: \"kubernetes.io/projected/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-kube-api-access-xwp6b\") pod \"community-operators-5rwgj\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.971247 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-utilities\") pod \"community-operators-5rwgj\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.971743 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-catalog-content\") pod \"community-operators-5rwgj\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.972009 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-utilities\") pod \"community-operators-5rwgj\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:11 crc kubenswrapper[4744]: I0106 15:55:11.999562 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwp6b\" (UniqueName: \"kubernetes.io/projected/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-kube-api-access-xwp6b\") pod \"community-operators-5rwgj\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:12 crc kubenswrapper[4744]: I0106 15:55:12.134034 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:12 crc kubenswrapper[4744]: I0106 15:55:12.665889 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rwgj"] Jan 06 15:55:12 crc kubenswrapper[4744]: I0106 15:55:12.786400 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rwgj" event={"ID":"bbfc106a-6af3-4740-b1df-fb52c97bcb8a","Type":"ContainerStarted","Data":"7d15b114939977fc913dede920bd5247871c2a9a9738e981ff44b5fa49339d84"} Jan 06 15:55:13 crc kubenswrapper[4744]: I0106 15:55:13.802297 4744 generic.go:334] "Generic (PLEG): container finished" podID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerID="842d39e270090a3823dd7a1ef212e8bb9f799117f19795f944a5ce40daa2bf6b" exitCode=0 Jan 06 15:55:13 crc kubenswrapper[4744]: I0106 15:55:13.802519 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rwgj" event={"ID":"bbfc106a-6af3-4740-b1df-fb52c97bcb8a","Type":"ContainerDied","Data":"842d39e270090a3823dd7a1ef212e8bb9f799117f19795f944a5ce40daa2bf6b"} Jan 06 15:55:15 crc kubenswrapper[4744]: I0106 15:55:15.978293 4744 generic.go:334] "Generic (PLEG): container finished" podID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerID="9d629550ca8cdb59994e7f6ea7846b8b659e7bf020f4ccb0185e413ea5854708" exitCode=0 Jan 06 15:55:15 crc kubenswrapper[4744]: I0106 15:55:15.978345 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rwgj" event={"ID":"bbfc106a-6af3-4740-b1df-fb52c97bcb8a","Type":"ContainerDied","Data":"9d629550ca8cdb59994e7f6ea7846b8b659e7bf020f4ccb0185e413ea5854708"} Jan 06 15:55:16 crc kubenswrapper[4744]: I0106 15:55:16.992622 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rwgj" event={"ID":"bbfc106a-6af3-4740-b1df-fb52c97bcb8a","Type":"ContainerStarted","Data":"b613d5d1c006f5a82b22ee5d311e6b70fb039c4d5feec816c6fa1a26f483885b"} Jan 06 15:55:17 crc kubenswrapper[4744]: I0106 15:55:17.021717 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5rwgj" podStartSLOduration=3.277382461 podStartE2EDuration="6.021695488s" podCreationTimestamp="2026-01-06 15:55:11 +0000 UTC" firstStartedPulling="2026-01-06 15:55:13.80575466 +0000 UTC m=+4710.433220988" lastFinishedPulling="2026-01-06 15:55:16.550067697 +0000 UTC m=+4713.177534015" observedRunningTime="2026-01-06 15:55:17.017334762 +0000 UTC m=+4713.644801090" watchObservedRunningTime="2026-01-06 15:55:17.021695488 +0000 UTC m=+4713.649161806" Jan 06 15:55:22 crc kubenswrapper[4744]: I0106 15:55:22.134482 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:22 crc kubenswrapper[4744]: I0106 15:55:22.135027 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:22 crc kubenswrapper[4744]: I0106 15:55:22.732828 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:23 crc kubenswrapper[4744]: I0106 15:55:23.164527 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:23 crc kubenswrapper[4744]: I0106 15:55:23.241855 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5rwgj"] Jan 06 15:55:25 crc kubenswrapper[4744]: I0106 15:55:25.115082 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5rwgj" podUID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerName="registry-server" containerID="cri-o://b613d5d1c006f5a82b22ee5d311e6b70fb039c4d5feec816c6fa1a26f483885b" gracePeriod=2 Jan 06 15:55:26 crc kubenswrapper[4744]: I0106 15:55:26.145882 4744 generic.go:334] "Generic (PLEG): container finished" podID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerID="b613d5d1c006f5a82b22ee5d311e6b70fb039c4d5feec816c6fa1a26f483885b" exitCode=0 Jan 06 15:55:26 crc kubenswrapper[4744]: I0106 15:55:26.146053 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rwgj" event={"ID":"bbfc106a-6af3-4740-b1df-fb52c97bcb8a","Type":"ContainerDied","Data":"b613d5d1c006f5a82b22ee5d311e6b70fb039c4d5feec816c6fa1a26f483885b"} Jan 06 15:55:26 crc kubenswrapper[4744]: I0106 15:55:26.972153 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.083226 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwp6b\" (UniqueName: \"kubernetes.io/projected/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-kube-api-access-xwp6b\") pod \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.083301 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-utilities\") pod \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.083656 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-catalog-content\") pod \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\" (UID: \"bbfc106a-6af3-4740-b1df-fb52c97bcb8a\") " Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.084590 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-utilities" (OuterVolumeSpecName: "utilities") pod "bbfc106a-6af3-4740-b1df-fb52c97bcb8a" (UID: "bbfc106a-6af3-4740-b1df-fb52c97bcb8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.096097 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-kube-api-access-xwp6b" (OuterVolumeSpecName: "kube-api-access-xwp6b") pod "bbfc106a-6af3-4740-b1df-fb52c97bcb8a" (UID: "bbfc106a-6af3-4740-b1df-fb52c97bcb8a"). InnerVolumeSpecName "kube-api-access-xwp6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.153111 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bbfc106a-6af3-4740-b1df-fb52c97bcb8a" (UID: "bbfc106a-6af3-4740-b1df-fb52c97bcb8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.164569 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rwgj" event={"ID":"bbfc106a-6af3-4740-b1df-fb52c97bcb8a","Type":"ContainerDied","Data":"7d15b114939977fc913dede920bd5247871c2a9a9738e981ff44b5fa49339d84"} Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.164644 4744 scope.go:117] "RemoveContainer" containerID="b613d5d1c006f5a82b22ee5d311e6b70fb039c4d5feec816c6fa1a26f483885b" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.164671 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rwgj" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.188308 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.188341 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwp6b\" (UniqueName: \"kubernetes.io/projected/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-kube-api-access-xwp6b\") on node \"crc\" DevicePath \"\"" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.188358 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bbfc106a-6af3-4740-b1df-fb52c97bcb8a-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.202316 4744 scope.go:117] "RemoveContainer" containerID="9d629550ca8cdb59994e7f6ea7846b8b659e7bf020f4ccb0185e413ea5854708" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.209878 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5rwgj"] Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.229123 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5rwgj"] Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.240978 4744 scope.go:117] "RemoveContainer" containerID="842d39e270090a3823dd7a1ef212e8bb9f799117f19795f944a5ce40daa2bf6b" Jan 06 15:55:27 crc kubenswrapper[4744]: I0106 15:55:27.723621 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" path="/var/lib/kubelet/pods/bbfc106a-6af3-4740-b1df-fb52c97bcb8a/volumes" Jan 06 15:56:14 crc kubenswrapper[4744]: I0106 15:56:14.423810 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:56:14 crc kubenswrapper[4744]: I0106 15:56:14.424281 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:56:44 crc kubenswrapper[4744]: I0106 15:56:44.423422 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:56:44 crc kubenswrapper[4744]: I0106 15:56:44.423935 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:57:14 crc kubenswrapper[4744]: I0106 15:57:14.424354 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:57:14 crc kubenswrapper[4744]: I0106 15:57:14.425041 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 15:57:14 crc kubenswrapper[4744]: I0106 15:57:14.425108 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 15:57:14 crc kubenswrapper[4744]: I0106 15:57:14.426440 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"de1ccb05de866c492da3b9c727016ce800e411146c0daf1657febcb845d3ec8c"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 15:57:14 crc kubenswrapper[4744]: I0106 15:57:14.426535 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://de1ccb05de866c492da3b9c727016ce800e411146c0daf1657febcb845d3ec8c" gracePeriod=600 Jan 06 15:57:16 crc kubenswrapper[4744]: I0106 15:57:16.684694 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="de1ccb05de866c492da3b9c727016ce800e411146c0daf1657febcb845d3ec8c" exitCode=0 Jan 06 15:57:16 crc kubenswrapper[4744]: I0106 15:57:16.684795 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"de1ccb05de866c492da3b9c727016ce800e411146c0daf1657febcb845d3ec8c"} Jan 06 15:57:16 crc kubenswrapper[4744]: I0106 15:57:16.685201 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8"} Jan 06 15:57:16 crc kubenswrapper[4744]: I0106 15:57:16.685223 4744 scope.go:117] "RemoveContainer" containerID="c631214d42d2ffba21ab035cc69b72d723590e5c040c3b23e53f56a2c93e504a" Jan 06 15:59:44 crc kubenswrapper[4744]: I0106 15:59:44.424415 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 15:59:44 crc kubenswrapper[4744]: I0106 15:59:44.425074 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.174134 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn"] Jan 06 16:00:00 crc kubenswrapper[4744]: E0106 16:00:00.175331 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerName="extract-content" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.175351 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerName="extract-content" Jan 06 16:00:00 crc kubenswrapper[4744]: E0106 16:00:00.175378 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerName="registry-server" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.175386 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerName="registry-server" Jan 06 16:00:00 crc kubenswrapper[4744]: E0106 16:00:00.175416 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerName="extract-utilities" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.175427 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerName="extract-utilities" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.175832 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbfc106a-6af3-4740-b1df-fb52c97bcb8a" containerName="registry-server" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.177124 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.179670 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.180248 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.215287 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn"] Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.274517 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fbe5f318-143a-43b7-a393-d89f2a158dc5-secret-volume\") pod \"collect-profiles-29461920-bxwgn\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.275248 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxj4w\" (UniqueName: \"kubernetes.io/projected/fbe5f318-143a-43b7-a393-d89f2a158dc5-kube-api-access-lxj4w\") pod \"collect-profiles-29461920-bxwgn\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.275903 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fbe5f318-143a-43b7-a393-d89f2a158dc5-config-volume\") pod \"collect-profiles-29461920-bxwgn\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.378375 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fbe5f318-143a-43b7-a393-d89f2a158dc5-config-volume\") pod \"collect-profiles-29461920-bxwgn\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.378438 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fbe5f318-143a-43b7-a393-d89f2a158dc5-secret-volume\") pod \"collect-profiles-29461920-bxwgn\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.379352 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fbe5f318-143a-43b7-a393-d89f2a158dc5-config-volume\") pod \"collect-profiles-29461920-bxwgn\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.379562 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxj4w\" (UniqueName: \"kubernetes.io/projected/fbe5f318-143a-43b7-a393-d89f2a158dc5-kube-api-access-lxj4w\") pod \"collect-profiles-29461920-bxwgn\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.387477 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fbe5f318-143a-43b7-a393-d89f2a158dc5-secret-volume\") pod \"collect-profiles-29461920-bxwgn\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.400148 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxj4w\" (UniqueName: \"kubernetes.io/projected/fbe5f318-143a-43b7-a393-d89f2a158dc5-kube-api-access-lxj4w\") pod \"collect-profiles-29461920-bxwgn\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:00 crc kubenswrapper[4744]: I0106 16:00:00.507930 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:01 crc kubenswrapper[4744]: I0106 16:00:01.011597 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn"] Jan 06 16:00:01 crc kubenswrapper[4744]: W0106 16:00:01.764694 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfbe5f318_143a_43b7_a393_d89f2a158dc5.slice/crio-e5ed1dcb5f05e0c9b07fec4d6cf461da0063a4f57612733d58a15b5c6d682766 WatchSource:0}: Error finding container e5ed1dcb5f05e0c9b07fec4d6cf461da0063a4f57612733d58a15b5c6d682766: Status 404 returned error can't find the container with id e5ed1dcb5f05e0c9b07fec4d6cf461da0063a4f57612733d58a15b5c6d682766 Jan 06 16:00:01 crc kubenswrapper[4744]: I0106 16:00:01.929865 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" event={"ID":"fbe5f318-143a-43b7-a393-d89f2a158dc5","Type":"ContainerStarted","Data":"e5ed1dcb5f05e0c9b07fec4d6cf461da0063a4f57612733d58a15b5c6d682766"} Jan 06 16:00:02 crc kubenswrapper[4744]: I0106 16:00:02.964087 4744 generic.go:334] "Generic (PLEG): container finished" podID="fbe5f318-143a-43b7-a393-d89f2a158dc5" containerID="d8414fa5287447f737e917c03ac15bd9442d1a1ef267adc1f8823303281b6a48" exitCode=0 Jan 06 16:00:02 crc kubenswrapper[4744]: I0106 16:00:02.964429 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" event={"ID":"fbe5f318-143a-43b7-a393-d89f2a158dc5","Type":"ContainerDied","Data":"d8414fa5287447f737e917c03ac15bd9442d1a1ef267adc1f8823303281b6a48"} Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.436735 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.583786 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fbe5f318-143a-43b7-a393-d89f2a158dc5-config-volume\") pod \"fbe5f318-143a-43b7-a393-d89f2a158dc5\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.584042 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxj4w\" (UniqueName: \"kubernetes.io/projected/fbe5f318-143a-43b7-a393-d89f2a158dc5-kube-api-access-lxj4w\") pod \"fbe5f318-143a-43b7-a393-d89f2a158dc5\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.584071 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fbe5f318-143a-43b7-a393-d89f2a158dc5-secret-volume\") pod \"fbe5f318-143a-43b7-a393-d89f2a158dc5\" (UID: \"fbe5f318-143a-43b7-a393-d89f2a158dc5\") " Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.584618 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fbe5f318-143a-43b7-a393-d89f2a158dc5-config-volume" (OuterVolumeSpecName: "config-volume") pod "fbe5f318-143a-43b7-a393-d89f2a158dc5" (UID: "fbe5f318-143a-43b7-a393-d89f2a158dc5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.584844 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fbe5f318-143a-43b7-a393-d89f2a158dc5-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.590249 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbe5f318-143a-43b7-a393-d89f2a158dc5-kube-api-access-lxj4w" (OuterVolumeSpecName: "kube-api-access-lxj4w") pod "fbe5f318-143a-43b7-a393-d89f2a158dc5" (UID: "fbe5f318-143a-43b7-a393-d89f2a158dc5"). InnerVolumeSpecName "kube-api-access-lxj4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.596345 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fbe5f318-143a-43b7-a393-d89f2a158dc5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fbe5f318-143a-43b7-a393-d89f2a158dc5" (UID: "fbe5f318-143a-43b7-a393-d89f2a158dc5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.687059 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxj4w\" (UniqueName: \"kubernetes.io/projected/fbe5f318-143a-43b7-a393-d89f2a158dc5-kube-api-access-lxj4w\") on node \"crc\" DevicePath \"\"" Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.687090 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fbe5f318-143a-43b7-a393-d89f2a158dc5-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.985060 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" event={"ID":"fbe5f318-143a-43b7-a393-d89f2a158dc5","Type":"ContainerDied","Data":"e5ed1dcb5f05e0c9b07fec4d6cf461da0063a4f57612733d58a15b5c6d682766"} Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.985097 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5ed1dcb5f05e0c9b07fec4d6cf461da0063a4f57612733d58a15b5c6d682766" Jan 06 16:00:04 crc kubenswrapper[4744]: I0106 16:00:04.985212 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461920-bxwgn" Jan 06 16:00:05 crc kubenswrapper[4744]: I0106 16:00:05.528845 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs"] Jan 06 16:00:05 crc kubenswrapper[4744]: I0106 16:00:05.539422 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461875-z62cs"] Jan 06 16:00:05 crc kubenswrapper[4744]: I0106 16:00:05.726154 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c" path="/var/lib/kubelet/pods/f29fbe9e-c32f-4c07-8d49-9cbd6addcc5c/volumes" Jan 06 16:00:13 crc kubenswrapper[4744]: I0106 16:00:13.701881 4744 scope.go:117] "RemoveContainer" containerID="c5366a46dc52c1062676c4b450fa191f3b76a427236e1fff7cee7de78c703970" Jan 06 16:00:14 crc kubenswrapper[4744]: I0106 16:00:14.424293 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:00:14 crc kubenswrapper[4744]: I0106 16:00:14.424678 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:00:44 crc kubenswrapper[4744]: I0106 16:00:44.424419 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:00:44 crc kubenswrapper[4744]: I0106 16:00:44.425092 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:00:44 crc kubenswrapper[4744]: I0106 16:00:44.425197 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 16:00:44 crc kubenswrapper[4744]: I0106 16:00:44.426410 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 16:00:44 crc kubenswrapper[4744]: I0106 16:00:44.426520 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" gracePeriod=600 Jan 06 16:00:44 crc kubenswrapper[4744]: E0106 16:00:44.581512 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:00:45 crc kubenswrapper[4744]: I0106 16:00:45.469498 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" exitCode=0 Jan 06 16:00:45 crc kubenswrapper[4744]: I0106 16:00:45.469594 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8"} Jan 06 16:00:45 crc kubenswrapper[4744]: I0106 16:00:45.469832 4744 scope.go:117] "RemoveContainer" containerID="de1ccb05de866c492da3b9c727016ce800e411146c0daf1657febcb845d3ec8c" Jan 06 16:00:45 crc kubenswrapper[4744]: I0106 16:00:45.470636 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:00:45 crc kubenswrapper[4744]: E0106 16:00:45.470980 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:00:56 crc kubenswrapper[4744]: I0106 16:00:56.711306 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:00:56 crc kubenswrapper[4744]: E0106 16:00:56.712330 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.166922 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29461921-5vhv9"] Jan 06 16:01:00 crc kubenswrapper[4744]: E0106 16:01:00.168032 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbe5f318-143a-43b7-a393-d89f2a158dc5" containerName="collect-profiles" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.168046 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbe5f318-143a-43b7-a393-d89f2a158dc5" containerName="collect-profiles" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.168329 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbe5f318-143a-43b7-a393-d89f2a158dc5" containerName="collect-profiles" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.169067 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.195124 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29461921-5vhv9"] Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.273346 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnbxl\" (UniqueName: \"kubernetes.io/projected/2da36365-1bbd-45e9-bd2f-392f7b34b25d-kube-api-access-mnbxl\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.273387 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-combined-ca-bundle\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.273472 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-config-data\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.273553 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-fernet-keys\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.375223 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-combined-ca-bundle\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.375264 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnbxl\" (UniqueName: \"kubernetes.io/projected/2da36365-1bbd-45e9-bd2f-392f7b34b25d-kube-api-access-mnbxl\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.375343 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-config-data\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.375428 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-fernet-keys\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.383646 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-combined-ca-bundle\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.383961 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-fernet-keys\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.384380 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-config-data\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.403460 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnbxl\" (UniqueName: \"kubernetes.io/projected/2da36365-1bbd-45e9-bd2f-392f7b34b25d-kube-api-access-mnbxl\") pod \"keystone-cron-29461921-5vhv9\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.512946 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:00 crc kubenswrapper[4744]: I0106 16:01:00.962966 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29461921-5vhv9"] Jan 06 16:01:01 crc kubenswrapper[4744]: I0106 16:01:01.696528 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29461921-5vhv9" event={"ID":"2da36365-1bbd-45e9-bd2f-392f7b34b25d","Type":"ContainerStarted","Data":"9f50e7ebbd180c20a8794e99f42f89c151fac8afe065ba5f855f00fcf2f35177"} Jan 06 16:01:01 crc kubenswrapper[4744]: I0106 16:01:01.696886 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29461921-5vhv9" event={"ID":"2da36365-1bbd-45e9-bd2f-392f7b34b25d","Type":"ContainerStarted","Data":"a88864b597cab490bf5e7be056ba62c058d221b914589bf3787a363fdd2968bd"} Jan 06 16:01:01 crc kubenswrapper[4744]: I0106 16:01:01.715311 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29461921-5vhv9" podStartSLOduration=1.715294512 podStartE2EDuration="1.715294512s" podCreationTimestamp="2026-01-06 16:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 16:01:01.711778138 +0000 UTC m=+5058.339244476" watchObservedRunningTime="2026-01-06 16:01:01.715294512 +0000 UTC m=+5058.342760840" Jan 06 16:01:03 crc kubenswrapper[4744]: I0106 16:01:03.731426 4744 generic.go:334] "Generic (PLEG): container finished" podID="2da36365-1bbd-45e9-bd2f-392f7b34b25d" containerID="9f50e7ebbd180c20a8794e99f42f89c151fac8afe065ba5f855f00fcf2f35177" exitCode=0 Jan 06 16:01:03 crc kubenswrapper[4744]: I0106 16:01:03.735249 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29461921-5vhv9" event={"ID":"2da36365-1bbd-45e9-bd2f-392f7b34b25d","Type":"ContainerDied","Data":"9f50e7ebbd180c20a8794e99f42f89c151fac8afe065ba5f855f00fcf2f35177"} Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.185148 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.304864 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnbxl\" (UniqueName: \"kubernetes.io/projected/2da36365-1bbd-45e9-bd2f-392f7b34b25d-kube-api-access-mnbxl\") pod \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.304978 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-fernet-keys\") pod \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.305049 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-combined-ca-bundle\") pod \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.305428 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-config-data\") pod \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\" (UID: \"2da36365-1bbd-45e9-bd2f-392f7b34b25d\") " Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.315122 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2da36365-1bbd-45e9-bd2f-392f7b34b25d" (UID: "2da36365-1bbd-45e9-bd2f-392f7b34b25d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.317354 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2da36365-1bbd-45e9-bd2f-392f7b34b25d-kube-api-access-mnbxl" (OuterVolumeSpecName: "kube-api-access-mnbxl") pod "2da36365-1bbd-45e9-bd2f-392f7b34b25d" (UID: "2da36365-1bbd-45e9-bd2f-392f7b34b25d"). InnerVolumeSpecName "kube-api-access-mnbxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.350449 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2da36365-1bbd-45e9-bd2f-392f7b34b25d" (UID: "2da36365-1bbd-45e9-bd2f-392f7b34b25d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.392272 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-config-data" (OuterVolumeSpecName: "config-data") pod "2da36365-1bbd-45e9-bd2f-392f7b34b25d" (UID: "2da36365-1bbd-45e9-bd2f-392f7b34b25d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.408653 4744 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-config-data\") on node \"crc\" DevicePath \"\"" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.408885 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnbxl\" (UniqueName: \"kubernetes.io/projected/2da36365-1bbd-45e9-bd2f-392f7b34b25d-kube-api-access-mnbxl\") on node \"crc\" DevicePath \"\"" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.408987 4744 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.409064 4744 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2da36365-1bbd-45e9-bd2f-392f7b34b25d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.755970 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29461921-5vhv9" event={"ID":"2da36365-1bbd-45e9-bd2f-392f7b34b25d","Type":"ContainerDied","Data":"a88864b597cab490bf5e7be056ba62c058d221b914589bf3787a363fdd2968bd"} Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.756011 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a88864b597cab490bf5e7be056ba62c058d221b914589bf3787a363fdd2968bd" Jan 06 16:01:05 crc kubenswrapper[4744]: I0106 16:01:05.756240 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29461921-5vhv9" Jan 06 16:01:11 crc kubenswrapper[4744]: I0106 16:01:11.711690 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:01:11 crc kubenswrapper[4744]: E0106 16:01:11.712919 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:01:24 crc kubenswrapper[4744]: I0106 16:01:24.712207 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:01:24 crc kubenswrapper[4744]: E0106 16:01:24.713592 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:01:35 crc kubenswrapper[4744]: I0106 16:01:35.711485 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:01:35 crc kubenswrapper[4744]: E0106 16:01:35.712651 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:01:49 crc kubenswrapper[4744]: I0106 16:01:49.712322 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:01:49 crc kubenswrapper[4744]: E0106 16:01:49.714078 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:02:02 crc kubenswrapper[4744]: I0106 16:02:02.711619 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:02:02 crc kubenswrapper[4744]: E0106 16:02:02.712740 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:02:13 crc kubenswrapper[4744]: I0106 16:02:13.901038 4744 trace.go:236] Trace[1885747207]: "Calculate volume metrics of storage for pod minio-dev/minio" (06-Jan-2026 16:02:12.899) (total time: 1001ms): Jan 06 16:02:13 crc kubenswrapper[4744]: Trace[1885747207]: [1.00148706s] [1.00148706s] END Jan 06 16:02:17 crc kubenswrapper[4744]: I0106 16:02:17.711808 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:02:17 crc kubenswrapper[4744]: E0106 16:02:17.712904 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:02:31 crc kubenswrapper[4744]: I0106 16:02:31.712263 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:02:31 crc kubenswrapper[4744]: E0106 16:02:31.713657 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.682869 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tw9qn"] Jan 06 16:02:40 crc kubenswrapper[4744]: E0106 16:02:40.684334 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2da36365-1bbd-45e9-bd2f-392f7b34b25d" containerName="keystone-cron" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.684360 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2da36365-1bbd-45e9-bd2f-392f7b34b25d" containerName="keystone-cron" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.684802 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2da36365-1bbd-45e9-bd2f-392f7b34b25d" containerName="keystone-cron" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.687640 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.701022 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tw9qn"] Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.770270 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-catalog-content\") pod \"certified-operators-tw9qn\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.770332 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-utilities\") pod \"certified-operators-tw9qn\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.770461 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9nsp\" (UniqueName: \"kubernetes.io/projected/22de974b-fca5-4ddf-b704-2ec2da25edfd-kube-api-access-n9nsp\") pod \"certified-operators-tw9qn\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.872874 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9nsp\" (UniqueName: \"kubernetes.io/projected/22de974b-fca5-4ddf-b704-2ec2da25edfd-kube-api-access-n9nsp\") pod \"certified-operators-tw9qn\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.873048 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-catalog-content\") pod \"certified-operators-tw9qn\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.873097 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-utilities\") pod \"certified-operators-tw9qn\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.874046 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-catalog-content\") pod \"certified-operators-tw9qn\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:40 crc kubenswrapper[4744]: I0106 16:02:40.874096 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-utilities\") pod \"certified-operators-tw9qn\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:41 crc kubenswrapper[4744]: I0106 16:02:41.457274 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9nsp\" (UniqueName: \"kubernetes.io/projected/22de974b-fca5-4ddf-b704-2ec2da25edfd-kube-api-access-n9nsp\") pod \"certified-operators-tw9qn\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:41 crc kubenswrapper[4744]: I0106 16:02:41.614672 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:42 crc kubenswrapper[4744]: I0106 16:02:42.096151 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tw9qn"] Jan 06 16:02:42 crc kubenswrapper[4744]: W0106 16:02:42.120622 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22de974b_fca5_4ddf_b704_2ec2da25edfd.slice/crio-f9ee8e2f34191a1d84b988a467a0fabd5fb277d0b56941ff459b831c7413d92b WatchSource:0}: Error finding container f9ee8e2f34191a1d84b988a467a0fabd5fb277d0b56941ff459b831c7413d92b: Status 404 returned error can't find the container with id f9ee8e2f34191a1d84b988a467a0fabd5fb277d0b56941ff459b831c7413d92b Jan 06 16:02:43 crc kubenswrapper[4744]: I0106 16:02:43.102297 4744 generic.go:334] "Generic (PLEG): container finished" podID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerID="3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9" exitCode=0 Jan 06 16:02:43 crc kubenswrapper[4744]: I0106 16:02:43.102404 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tw9qn" event={"ID":"22de974b-fca5-4ddf-b704-2ec2da25edfd","Type":"ContainerDied","Data":"3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9"} Jan 06 16:02:43 crc kubenswrapper[4744]: I0106 16:02:43.103029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tw9qn" event={"ID":"22de974b-fca5-4ddf-b704-2ec2da25edfd","Type":"ContainerStarted","Data":"f9ee8e2f34191a1d84b988a467a0fabd5fb277d0b56941ff459b831c7413d92b"} Jan 06 16:02:43 crc kubenswrapper[4744]: I0106 16:02:43.107296 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 16:02:44 crc kubenswrapper[4744]: I0106 16:02:44.120986 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tw9qn" event={"ID":"22de974b-fca5-4ddf-b704-2ec2da25edfd","Type":"ContainerStarted","Data":"8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3"} Jan 06 16:02:44 crc kubenswrapper[4744]: I0106 16:02:44.711947 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:02:44 crc kubenswrapper[4744]: E0106 16:02:44.712928 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:02:46 crc kubenswrapper[4744]: I0106 16:02:46.153022 4744 generic.go:334] "Generic (PLEG): container finished" podID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerID="8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3" exitCode=0 Jan 06 16:02:46 crc kubenswrapper[4744]: I0106 16:02:46.153133 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tw9qn" event={"ID":"22de974b-fca5-4ddf-b704-2ec2da25edfd","Type":"ContainerDied","Data":"8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3"} Jan 06 16:02:47 crc kubenswrapper[4744]: I0106 16:02:47.166641 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tw9qn" event={"ID":"22de974b-fca5-4ddf-b704-2ec2da25edfd","Type":"ContainerStarted","Data":"94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91"} Jan 06 16:02:47 crc kubenswrapper[4744]: I0106 16:02:47.192676 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tw9qn" podStartSLOduration=3.490773738 podStartE2EDuration="7.19266092s" podCreationTimestamp="2026-01-06 16:02:40 +0000 UTC" firstStartedPulling="2026-01-06 16:02:43.106654218 +0000 UTC m=+5159.734120566" lastFinishedPulling="2026-01-06 16:02:46.80854142 +0000 UTC m=+5163.436007748" observedRunningTime="2026-01-06 16:02:47.183639338 +0000 UTC m=+5163.811105656" watchObservedRunningTime="2026-01-06 16:02:47.19266092 +0000 UTC m=+5163.820127238" Jan 06 16:02:51 crc kubenswrapper[4744]: I0106 16:02:51.615802 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:51 crc kubenswrapper[4744]: I0106 16:02:51.616317 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:51 crc kubenswrapper[4744]: I0106 16:02:51.927734 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:52 crc kubenswrapper[4744]: I0106 16:02:52.301858 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:52 crc kubenswrapper[4744]: I0106 16:02:52.375671 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tw9qn"] Jan 06 16:02:54 crc kubenswrapper[4744]: I0106 16:02:54.243505 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tw9qn" podUID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerName="registry-server" containerID="cri-o://94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91" gracePeriod=2 Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.004525 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.070879 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-catalog-content\") pod \"22de974b-fca5-4ddf-b704-2ec2da25edfd\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.071003 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9nsp\" (UniqueName: \"kubernetes.io/projected/22de974b-fca5-4ddf-b704-2ec2da25edfd-kube-api-access-n9nsp\") pod \"22de974b-fca5-4ddf-b704-2ec2da25edfd\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.071087 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-utilities\") pod \"22de974b-fca5-4ddf-b704-2ec2da25edfd\" (UID: \"22de974b-fca5-4ddf-b704-2ec2da25edfd\") " Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.072596 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-utilities" (OuterVolumeSpecName: "utilities") pod "22de974b-fca5-4ddf-b704-2ec2da25edfd" (UID: "22de974b-fca5-4ddf-b704-2ec2da25edfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.078535 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22de974b-fca5-4ddf-b704-2ec2da25edfd-kube-api-access-n9nsp" (OuterVolumeSpecName: "kube-api-access-n9nsp") pod "22de974b-fca5-4ddf-b704-2ec2da25edfd" (UID: "22de974b-fca5-4ddf-b704-2ec2da25edfd"). InnerVolumeSpecName "kube-api-access-n9nsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.131513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22de974b-fca5-4ddf-b704-2ec2da25edfd" (UID: "22de974b-fca5-4ddf-b704-2ec2da25edfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.174042 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.174071 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9nsp\" (UniqueName: \"kubernetes.io/projected/22de974b-fca5-4ddf-b704-2ec2da25edfd-kube-api-access-n9nsp\") on node \"crc\" DevicePath \"\"" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.174082 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22de974b-fca5-4ddf-b704-2ec2da25edfd-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.254656 4744 generic.go:334] "Generic (PLEG): container finished" podID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerID="94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91" exitCode=0 Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.254694 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tw9qn" event={"ID":"22de974b-fca5-4ddf-b704-2ec2da25edfd","Type":"ContainerDied","Data":"94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91"} Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.254722 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tw9qn" event={"ID":"22de974b-fca5-4ddf-b704-2ec2da25edfd","Type":"ContainerDied","Data":"f9ee8e2f34191a1d84b988a467a0fabd5fb277d0b56941ff459b831c7413d92b"} Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.254742 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tw9qn" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.254746 4744 scope.go:117] "RemoveContainer" containerID="94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.287022 4744 scope.go:117] "RemoveContainer" containerID="8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.304909 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tw9qn"] Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.318633 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tw9qn"] Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.340743 4744 scope.go:117] "RemoveContainer" containerID="3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.405019 4744 scope.go:117] "RemoveContainer" containerID="94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91" Jan 06 16:02:55 crc kubenswrapper[4744]: E0106 16:02:55.405474 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91\": container with ID starting with 94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91 not found: ID does not exist" containerID="94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.405542 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91"} err="failed to get container status \"94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91\": rpc error: code = NotFound desc = could not find container \"94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91\": container with ID starting with 94e77030e391bbd8e2f1abdc1dff0cd0a88c4424f2fa478c26c3427d1bbd3e91 not found: ID does not exist" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.405577 4744 scope.go:117] "RemoveContainer" containerID="8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3" Jan 06 16:02:55 crc kubenswrapper[4744]: E0106 16:02:55.406084 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3\": container with ID starting with 8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3 not found: ID does not exist" containerID="8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.406363 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3"} err="failed to get container status \"8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3\": rpc error: code = NotFound desc = could not find container \"8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3\": container with ID starting with 8f23e1e585a7b93151fc03271ac787c4a509ca59aa3ec34fa00bbb44ce8180b3 not found: ID does not exist" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.406551 4744 scope.go:117] "RemoveContainer" containerID="3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9" Jan 06 16:02:55 crc kubenswrapper[4744]: E0106 16:02:55.407123 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9\": container with ID starting with 3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9 not found: ID does not exist" containerID="3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.407258 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9"} err="failed to get container status \"3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9\": rpc error: code = NotFound desc = could not find container \"3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9\": container with ID starting with 3f62d1c9b12849eb179d22b59b428d3ac3d94056bf37063e48327a39fe67aaf9 not found: ID does not exist" Jan 06 16:02:55 crc kubenswrapper[4744]: E0106 16:02:55.494724 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22de974b_fca5_4ddf_b704_2ec2da25edfd.slice/crio-f9ee8e2f34191a1d84b988a467a0fabd5fb277d0b56941ff459b831c7413d92b\": RecentStats: unable to find data in memory cache]" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.711682 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:02:55 crc kubenswrapper[4744]: E0106 16:02:55.711908 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:02:55 crc kubenswrapper[4744]: I0106 16:02:55.738782 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22de974b-fca5-4ddf-b704-2ec2da25edfd" path="/var/lib/kubelet/pods/22de974b-fca5-4ddf-b704-2ec2da25edfd/volumes" Jan 06 16:03:07 crc kubenswrapper[4744]: I0106 16:03:07.711606 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:03:07 crc kubenswrapper[4744]: E0106 16:03:07.712391 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:03:18 crc kubenswrapper[4744]: I0106 16:03:18.712367 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:03:18 crc kubenswrapper[4744]: E0106 16:03:18.713583 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.015813 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5g8b7"] Jan 06 16:03:21 crc kubenswrapper[4744]: E0106 16:03:21.018063 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerName="extract-utilities" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.018089 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerName="extract-utilities" Jan 06 16:03:21 crc kubenswrapper[4744]: E0106 16:03:21.018157 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerName="extract-content" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.018184 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerName="extract-content" Jan 06 16:03:21 crc kubenswrapper[4744]: E0106 16:03:21.018241 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerName="registry-server" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.019107 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerName="registry-server" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.019719 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="22de974b-fca5-4ddf-b704-2ec2da25edfd" containerName="registry-server" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.025016 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.043832 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g8b7"] Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.121619 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpqkj\" (UniqueName: \"kubernetes.io/projected/28fd5185-1662-4f0a-a957-0de3734b8fa1-kube-api-access-wpqkj\") pod \"redhat-marketplace-5g8b7\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.121780 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-utilities\") pod \"redhat-marketplace-5g8b7\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.121828 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-catalog-content\") pod \"redhat-marketplace-5g8b7\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.224067 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-utilities\") pod \"redhat-marketplace-5g8b7\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.224194 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-catalog-content\") pod \"redhat-marketplace-5g8b7\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.224346 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpqkj\" (UniqueName: \"kubernetes.io/projected/28fd5185-1662-4f0a-a957-0de3734b8fa1-kube-api-access-wpqkj\") pod \"redhat-marketplace-5g8b7\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.224763 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-utilities\") pod \"redhat-marketplace-5g8b7\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.224794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-catalog-content\") pod \"redhat-marketplace-5g8b7\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.755995 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpqkj\" (UniqueName: \"kubernetes.io/projected/28fd5185-1662-4f0a-a957-0de3734b8fa1-kube-api-access-wpqkj\") pod \"redhat-marketplace-5g8b7\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:21 crc kubenswrapper[4744]: I0106 16:03:21.959084 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:22 crc kubenswrapper[4744]: I0106 16:03:22.450184 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g8b7"] Jan 06 16:03:22 crc kubenswrapper[4744]: I0106 16:03:22.654716 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g8b7" event={"ID":"28fd5185-1662-4f0a-a957-0de3734b8fa1","Type":"ContainerStarted","Data":"b77da860b9ea03fc7e4af12a1e61e30bb87b8860f2e510bf57aa9cefe4cc477e"} Jan 06 16:03:23 crc kubenswrapper[4744]: I0106 16:03:23.671901 4744 generic.go:334] "Generic (PLEG): container finished" podID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerID="7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5" exitCode=0 Jan 06 16:03:23 crc kubenswrapper[4744]: I0106 16:03:23.672007 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g8b7" event={"ID":"28fd5185-1662-4f0a-a957-0de3734b8fa1","Type":"ContainerDied","Data":"7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5"} Jan 06 16:03:25 crc kubenswrapper[4744]: I0106 16:03:25.707720 4744 generic.go:334] "Generic (PLEG): container finished" podID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerID="6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038" exitCode=0 Jan 06 16:03:25 crc kubenswrapper[4744]: I0106 16:03:25.707800 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g8b7" event={"ID":"28fd5185-1662-4f0a-a957-0de3734b8fa1","Type":"ContainerDied","Data":"6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038"} Jan 06 16:03:26 crc kubenswrapper[4744]: I0106 16:03:26.731342 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g8b7" event={"ID":"28fd5185-1662-4f0a-a957-0de3734b8fa1","Type":"ContainerStarted","Data":"e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd"} Jan 06 16:03:26 crc kubenswrapper[4744]: I0106 16:03:26.770496 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5g8b7" podStartSLOduration=4.291020261 podStartE2EDuration="6.77046987s" podCreationTimestamp="2026-01-06 16:03:20 +0000 UTC" firstStartedPulling="2026-01-06 16:03:23.674520583 +0000 UTC m=+5200.301986891" lastFinishedPulling="2026-01-06 16:03:26.153970142 +0000 UTC m=+5202.781436500" observedRunningTime="2026-01-06 16:03:26.760587556 +0000 UTC m=+5203.388053914" watchObservedRunningTime="2026-01-06 16:03:26.77046987 +0000 UTC m=+5203.397936218" Jan 06 16:03:31 crc kubenswrapper[4744]: I0106 16:03:31.959328 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:31 crc kubenswrapper[4744]: I0106 16:03:31.960082 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:32 crc kubenswrapper[4744]: I0106 16:03:32.055966 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:32 crc kubenswrapper[4744]: I0106 16:03:32.897212 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:32 crc kubenswrapper[4744]: I0106 16:03:32.953434 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g8b7"] Jan 06 16:03:33 crc kubenswrapper[4744]: I0106 16:03:33.729517 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:03:33 crc kubenswrapper[4744]: E0106 16:03:33.731507 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:03:34 crc kubenswrapper[4744]: I0106 16:03:34.848607 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5g8b7" podUID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerName="registry-server" containerID="cri-o://e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd" gracePeriod=2 Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.426878 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.539409 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpqkj\" (UniqueName: \"kubernetes.io/projected/28fd5185-1662-4f0a-a957-0de3734b8fa1-kube-api-access-wpqkj\") pod \"28fd5185-1662-4f0a-a957-0de3734b8fa1\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.539481 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-catalog-content\") pod \"28fd5185-1662-4f0a-a957-0de3734b8fa1\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.539542 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-utilities\") pod \"28fd5185-1662-4f0a-a957-0de3734b8fa1\" (UID: \"28fd5185-1662-4f0a-a957-0de3734b8fa1\") " Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.540513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-utilities" (OuterVolumeSpecName: "utilities") pod "28fd5185-1662-4f0a-a957-0de3734b8fa1" (UID: "28fd5185-1662-4f0a-a957-0de3734b8fa1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.550981 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28fd5185-1662-4f0a-a957-0de3734b8fa1-kube-api-access-wpqkj" (OuterVolumeSpecName: "kube-api-access-wpqkj") pod "28fd5185-1662-4f0a-a957-0de3734b8fa1" (UID: "28fd5185-1662-4f0a-a957-0de3734b8fa1"). InnerVolumeSpecName "kube-api-access-wpqkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.576066 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28fd5185-1662-4f0a-a957-0de3734b8fa1" (UID: "28fd5185-1662-4f0a-a957-0de3734b8fa1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.642563 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpqkj\" (UniqueName: \"kubernetes.io/projected/28fd5185-1662-4f0a-a957-0de3734b8fa1-kube-api-access-wpqkj\") on node \"crc\" DevicePath \"\"" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.642601 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.642616 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28fd5185-1662-4f0a-a957-0de3734b8fa1-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.863874 4744 generic.go:334] "Generic (PLEG): container finished" podID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerID="e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd" exitCode=0 Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.863956 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5g8b7" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.863949 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g8b7" event={"ID":"28fd5185-1662-4f0a-a957-0de3734b8fa1","Type":"ContainerDied","Data":"e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd"} Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.864117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5g8b7" event={"ID":"28fd5185-1662-4f0a-a957-0de3734b8fa1","Type":"ContainerDied","Data":"b77da860b9ea03fc7e4af12a1e61e30bb87b8860f2e510bf57aa9cefe4cc477e"} Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.864148 4744 scope.go:117] "RemoveContainer" containerID="e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.904961 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g8b7"] Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.913662 4744 scope.go:117] "RemoveContainer" containerID="6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038" Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.916835 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5g8b7"] Jan 06 16:03:35 crc kubenswrapper[4744]: I0106 16:03:35.956244 4744 scope.go:117] "RemoveContainer" containerID="7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5" Jan 06 16:03:36 crc kubenswrapper[4744]: I0106 16:03:36.011713 4744 scope.go:117] "RemoveContainer" containerID="e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd" Jan 06 16:03:36 crc kubenswrapper[4744]: E0106 16:03:36.012481 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd\": container with ID starting with e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd not found: ID does not exist" containerID="e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd" Jan 06 16:03:36 crc kubenswrapper[4744]: I0106 16:03:36.012534 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd"} err="failed to get container status \"e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd\": rpc error: code = NotFound desc = could not find container \"e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd\": container with ID starting with e9697c4b1a6687c489e57ce3eee39b336e250c4f476162d5a256df91854fdbbd not found: ID does not exist" Jan 06 16:03:36 crc kubenswrapper[4744]: I0106 16:03:36.012568 4744 scope.go:117] "RemoveContainer" containerID="6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038" Jan 06 16:03:36 crc kubenswrapper[4744]: E0106 16:03:36.013062 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038\": container with ID starting with 6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038 not found: ID does not exist" containerID="6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038" Jan 06 16:03:36 crc kubenswrapper[4744]: I0106 16:03:36.013128 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038"} err="failed to get container status \"6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038\": rpc error: code = NotFound desc = could not find container \"6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038\": container with ID starting with 6f2ed8b35ff637c656feef09dbc4f75572954b0a68028daf35743704856b0038 not found: ID does not exist" Jan 06 16:03:36 crc kubenswrapper[4744]: I0106 16:03:36.013197 4744 scope.go:117] "RemoveContainer" containerID="7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5" Jan 06 16:03:36 crc kubenswrapper[4744]: E0106 16:03:36.013623 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5\": container with ID starting with 7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5 not found: ID does not exist" containerID="7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5" Jan 06 16:03:36 crc kubenswrapper[4744]: I0106 16:03:36.013669 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5"} err="failed to get container status \"7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5\": rpc error: code = NotFound desc = could not find container \"7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5\": container with ID starting with 7e60dc8e589728e026f7256ed268a22221017d39f66db32d97984324bc0c1db5 not found: ID does not exist" Jan 06 16:03:37 crc kubenswrapper[4744]: I0106 16:03:37.737971 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28fd5185-1662-4f0a-a957-0de3734b8fa1" path="/var/lib/kubelet/pods/28fd5185-1662-4f0a-a957-0de3734b8fa1/volumes" Jan 06 16:03:48 crc kubenswrapper[4744]: I0106 16:03:48.712145 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:03:48 crc kubenswrapper[4744]: E0106 16:03:48.713329 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:03:59 crc kubenswrapper[4744]: I0106 16:03:59.711509 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:03:59 crc kubenswrapper[4744]: E0106 16:03:59.712458 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:04:14 crc kubenswrapper[4744]: I0106 16:04:14.714218 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:04:14 crc kubenswrapper[4744]: E0106 16:04:14.715363 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:04:28 crc kubenswrapper[4744]: I0106 16:04:28.712133 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:04:28 crc kubenswrapper[4744]: E0106 16:04:28.713455 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:04:42 crc kubenswrapper[4744]: I0106 16:04:42.712248 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:04:42 crc kubenswrapper[4744]: E0106 16:04:42.713542 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:04:57 crc kubenswrapper[4744]: I0106 16:04:57.712568 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:04:57 crc kubenswrapper[4744]: E0106 16:04:57.713729 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:05:12 crc kubenswrapper[4744]: I0106 16:05:12.149209 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:05:12 crc kubenswrapper[4744]: E0106 16:05:12.149954 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:05:22 crc kubenswrapper[4744]: I0106 16:05:22.711886 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:05:22 crc kubenswrapper[4744]: E0106 16:05:22.713050 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:05:35 crc kubenswrapper[4744]: I0106 16:05:35.711744 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:05:35 crc kubenswrapper[4744]: E0106 16:05:35.712735 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:05:49 crc kubenswrapper[4744]: I0106 16:05:49.711939 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:05:50 crc kubenswrapper[4744]: I0106 16:05:50.714857 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"1fce60955246c02a8c5fa277cc49d89bc99389098884f65539257aa5968cb082"} Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.046697 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rqrl2"] Jan 06 16:06:15 crc kubenswrapper[4744]: E0106 16:06:15.048106 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerName="extract-content" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.048131 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerName="extract-content" Jan 06 16:06:15 crc kubenswrapper[4744]: E0106 16:06:15.048248 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerName="registry-server" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.048262 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerName="registry-server" Jan 06 16:06:15 crc kubenswrapper[4744]: E0106 16:06:15.048294 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerName="extract-utilities" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.048307 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerName="extract-utilities" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.048701 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="28fd5185-1662-4f0a-a957-0de3734b8fa1" containerName="registry-server" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.054200 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.077630 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqrl2"] Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.158642 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-utilities\") pod \"community-operators-rqrl2\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.158738 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plmt8\" (UniqueName: \"kubernetes.io/projected/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-kube-api-access-plmt8\") pod \"community-operators-rqrl2\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.158800 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-catalog-content\") pod \"community-operators-rqrl2\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.261245 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-utilities\") pod \"community-operators-rqrl2\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.261368 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plmt8\" (UniqueName: \"kubernetes.io/projected/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-kube-api-access-plmt8\") pod \"community-operators-rqrl2\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.261447 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-catalog-content\") pod \"community-operators-rqrl2\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.262190 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-catalog-content\") pod \"community-operators-rqrl2\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.262186 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-utilities\") pod \"community-operators-rqrl2\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.283978 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plmt8\" (UniqueName: \"kubernetes.io/projected/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-kube-api-access-plmt8\") pod \"community-operators-rqrl2\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.384048 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:15 crc kubenswrapper[4744]: I0106 16:06:15.921338 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rqrl2"] Jan 06 16:06:15 crc kubenswrapper[4744]: W0106 16:06:15.932027 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae9aab4c_f38c_4c8b_bd37_ecde31b14613.slice/crio-03e9e91cd1bf721cfa7d8613a73b0aa90de541806c0799d46e569de10e161f0e WatchSource:0}: Error finding container 03e9e91cd1bf721cfa7d8613a73b0aa90de541806c0799d46e569de10e161f0e: Status 404 returned error can't find the container with id 03e9e91cd1bf721cfa7d8613a73b0aa90de541806c0799d46e569de10e161f0e Jan 06 16:06:16 crc kubenswrapper[4744]: I0106 16:06:16.113480 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqrl2" event={"ID":"ae9aab4c-f38c-4c8b-bd37-ecde31b14613","Type":"ContainerStarted","Data":"03e9e91cd1bf721cfa7d8613a73b0aa90de541806c0799d46e569de10e161f0e"} Jan 06 16:06:17 crc kubenswrapper[4744]: I0106 16:06:17.131835 4744 generic.go:334] "Generic (PLEG): container finished" podID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerID="19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e" exitCode=0 Jan 06 16:06:17 crc kubenswrapper[4744]: I0106 16:06:17.131905 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqrl2" event={"ID":"ae9aab4c-f38c-4c8b-bd37-ecde31b14613","Type":"ContainerDied","Data":"19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e"} Jan 06 16:06:19 crc kubenswrapper[4744]: I0106 16:06:19.171858 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqrl2" event={"ID":"ae9aab4c-f38c-4c8b-bd37-ecde31b14613","Type":"ContainerStarted","Data":"52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0"} Jan 06 16:06:20 crc kubenswrapper[4744]: I0106 16:06:20.186413 4744 generic.go:334] "Generic (PLEG): container finished" podID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerID="52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0" exitCode=0 Jan 06 16:06:20 crc kubenswrapper[4744]: I0106 16:06:20.186696 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqrl2" event={"ID":"ae9aab4c-f38c-4c8b-bd37-ecde31b14613","Type":"ContainerDied","Data":"52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0"} Jan 06 16:06:22 crc kubenswrapper[4744]: I0106 16:06:22.217803 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqrl2" event={"ID":"ae9aab4c-f38c-4c8b-bd37-ecde31b14613","Type":"ContainerStarted","Data":"4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69"} Jan 06 16:06:22 crc kubenswrapper[4744]: I0106 16:06:22.251824 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rqrl2" podStartSLOduration=3.7167895570000002 podStartE2EDuration="7.251804024s" podCreationTimestamp="2026-01-06 16:06:15 +0000 UTC" firstStartedPulling="2026-01-06 16:06:17.136068927 +0000 UTC m=+5373.763535285" lastFinishedPulling="2026-01-06 16:06:20.671083434 +0000 UTC m=+5377.298549752" observedRunningTime="2026-01-06 16:06:22.244372355 +0000 UTC m=+5378.871838703" watchObservedRunningTime="2026-01-06 16:06:22.251804024 +0000 UTC m=+5378.879270352" Jan 06 16:06:25 crc kubenswrapper[4744]: I0106 16:06:25.384899 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:25 crc kubenswrapper[4744]: I0106 16:06:25.385627 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:25 crc kubenswrapper[4744]: I0106 16:06:25.461374 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:26 crc kubenswrapper[4744]: I0106 16:06:26.361691 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:26 crc kubenswrapper[4744]: I0106 16:06:26.436652 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqrl2"] Jan 06 16:06:28 crc kubenswrapper[4744]: I0106 16:06:28.312895 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rqrl2" podUID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerName="registry-server" containerID="cri-o://4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69" gracePeriod=2 Jan 06 16:06:28 crc kubenswrapper[4744]: I0106 16:06:28.900864 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.040932 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plmt8\" (UniqueName: \"kubernetes.io/projected/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-kube-api-access-plmt8\") pod \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.041026 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-catalog-content\") pod \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.041062 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-utilities\") pod \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\" (UID: \"ae9aab4c-f38c-4c8b-bd37-ecde31b14613\") " Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.042475 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-utilities" (OuterVolumeSpecName: "utilities") pod "ae9aab4c-f38c-4c8b-bd37-ecde31b14613" (UID: "ae9aab4c-f38c-4c8b-bd37-ecde31b14613"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.049697 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-kube-api-access-plmt8" (OuterVolumeSpecName: "kube-api-access-plmt8") pod "ae9aab4c-f38c-4c8b-bd37-ecde31b14613" (UID: "ae9aab4c-f38c-4c8b-bd37-ecde31b14613"). InnerVolumeSpecName "kube-api-access-plmt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.096111 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae9aab4c-f38c-4c8b-bd37-ecde31b14613" (UID: "ae9aab4c-f38c-4c8b-bd37-ecde31b14613"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.143706 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.143739 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.143749 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plmt8\" (UniqueName: \"kubernetes.io/projected/ae9aab4c-f38c-4c8b-bd37-ecde31b14613-kube-api-access-plmt8\") on node \"crc\" DevicePath \"\"" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.332192 4744 generic.go:334] "Generic (PLEG): container finished" podID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerID="4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69" exitCode=0 Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.332254 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqrl2" event={"ID":"ae9aab4c-f38c-4c8b-bd37-ecde31b14613","Type":"ContainerDied","Data":"4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69"} Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.332297 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rqrl2" event={"ID":"ae9aab4c-f38c-4c8b-bd37-ecde31b14613","Type":"ContainerDied","Data":"03e9e91cd1bf721cfa7d8613a73b0aa90de541806c0799d46e569de10e161f0e"} Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.332329 4744 scope.go:117] "RemoveContainer" containerID="4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.332585 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rqrl2" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.370716 4744 scope.go:117] "RemoveContainer" containerID="52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.393089 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rqrl2"] Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.403636 4744 scope.go:117] "RemoveContainer" containerID="19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.410982 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rqrl2"] Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.467266 4744 scope.go:117] "RemoveContainer" containerID="4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69" Jan 06 16:06:29 crc kubenswrapper[4744]: E0106 16:06:29.467869 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69\": container with ID starting with 4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69 not found: ID does not exist" containerID="4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.467918 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69"} err="failed to get container status \"4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69\": rpc error: code = NotFound desc = could not find container \"4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69\": container with ID starting with 4c839d5425fc859b5f1592816c2a853d5b571f42196001526e509f2354610d69 not found: ID does not exist" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.467953 4744 scope.go:117] "RemoveContainer" containerID="52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0" Jan 06 16:06:29 crc kubenswrapper[4744]: E0106 16:06:29.468415 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0\": container with ID starting with 52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0 not found: ID does not exist" containerID="52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.468453 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0"} err="failed to get container status \"52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0\": rpc error: code = NotFound desc = could not find container \"52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0\": container with ID starting with 52ad54f868abf15f3c99a9bee199a37237f374d1f83b8d6cb1f02932ad77fff0 not found: ID does not exist" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.468480 4744 scope.go:117] "RemoveContainer" containerID="19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e" Jan 06 16:06:29 crc kubenswrapper[4744]: E0106 16:06:29.468938 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e\": container with ID starting with 19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e not found: ID does not exist" containerID="19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.468973 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e"} err="failed to get container status \"19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e\": rpc error: code = NotFound desc = could not find container \"19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e\": container with ID starting with 19f01e60aa8703389fd5808d9bbbe1a4e3a96a5ae1e8ed6241e5ff5e92321f8e not found: ID does not exist" Jan 06 16:06:29 crc kubenswrapper[4744]: I0106 16:06:29.727752 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" path="/var/lib/kubelet/pods/ae9aab4c-f38c-4c8b-bd37-ecde31b14613/volumes" Jan 06 16:08:10 crc kubenswrapper[4744]: E0106 16:08:10.859757 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.22:49290->38.102.83.22:45165: write tcp 38.102.83.22:49290->38.102.83.22:45165: write: broken pipe Jan 06 16:08:14 crc kubenswrapper[4744]: I0106 16:08:14.424682 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:08:14 crc kubenswrapper[4744]: I0106 16:08:14.425372 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:08:32 crc kubenswrapper[4744]: E0106 16:08:32.815045 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.22:41854->38.102.83.22:45165: write tcp 38.102.83.22:41854->38.102.83.22:45165: write: broken pipe Jan 06 16:08:44 crc kubenswrapper[4744]: I0106 16:08:44.424309 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:08:44 crc kubenswrapper[4744]: I0106 16:08:44.424772 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:09:14 crc kubenswrapper[4744]: I0106 16:09:14.423925 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:09:14 crc kubenswrapper[4744]: I0106 16:09:14.424627 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:09:14 crc kubenswrapper[4744]: I0106 16:09:14.424700 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 16:09:14 crc kubenswrapper[4744]: I0106 16:09:14.425999 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1fce60955246c02a8c5fa277cc49d89bc99389098884f65539257aa5968cb082"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 16:09:14 crc kubenswrapper[4744]: I0106 16:09:14.426113 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://1fce60955246c02a8c5fa277cc49d89bc99389098884f65539257aa5968cb082" gracePeriod=600 Jan 06 16:09:14 crc kubenswrapper[4744]: I0106 16:09:14.664797 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="1fce60955246c02a8c5fa277cc49d89bc99389098884f65539257aa5968cb082" exitCode=0 Jan 06 16:09:14 crc kubenswrapper[4744]: I0106 16:09:14.664868 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"1fce60955246c02a8c5fa277cc49d89bc99389098884f65539257aa5968cb082"} Jan 06 16:09:14 crc kubenswrapper[4744]: I0106 16:09:14.664955 4744 scope.go:117] "RemoveContainer" containerID="e4df0fffa90ba6fe6c1543adec0f1db0ba81cf1926b51ef3006cd035d55725a8" Jan 06 16:09:15 crc kubenswrapper[4744]: I0106 16:09:15.697741 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb"} Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.770185 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ftdpc"] Jan 06 16:09:58 crc kubenswrapper[4744]: E0106 16:09:58.772288 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerName="registry-server" Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.772376 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerName="registry-server" Jan 06 16:09:58 crc kubenswrapper[4744]: E0106 16:09:58.772483 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerName="extract-content" Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.772540 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerName="extract-content" Jan 06 16:09:58 crc kubenswrapper[4744]: E0106 16:09:58.772618 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerName="extract-utilities" Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.772678 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerName="extract-utilities" Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.772984 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae9aab4c-f38c-4c8b-bd37-ecde31b14613" containerName="registry-server" Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.774885 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.799212 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ftdpc"] Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.978960 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mm4v\" (UniqueName: \"kubernetes.io/projected/3d08e4f4-7126-49f8-898d-4441f23b2245-kube-api-access-7mm4v\") pod \"redhat-operators-ftdpc\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.979219 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-catalog-content\") pod \"redhat-operators-ftdpc\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:58 crc kubenswrapper[4744]: I0106 16:09:58.979389 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-utilities\") pod \"redhat-operators-ftdpc\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:59 crc kubenswrapper[4744]: I0106 16:09:59.086464 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-utilities\") pod \"redhat-operators-ftdpc\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:59 crc kubenswrapper[4744]: I0106 16:09:59.086725 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mm4v\" (UniqueName: \"kubernetes.io/projected/3d08e4f4-7126-49f8-898d-4441f23b2245-kube-api-access-7mm4v\") pod \"redhat-operators-ftdpc\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:59 crc kubenswrapper[4744]: I0106 16:09:59.086798 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-catalog-content\") pod \"redhat-operators-ftdpc\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:59 crc kubenswrapper[4744]: I0106 16:09:59.086930 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-utilities\") pod \"redhat-operators-ftdpc\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:59 crc kubenswrapper[4744]: I0106 16:09:59.087633 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-catalog-content\") pod \"redhat-operators-ftdpc\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:59 crc kubenswrapper[4744]: I0106 16:09:59.118106 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mm4v\" (UniqueName: \"kubernetes.io/projected/3d08e4f4-7126-49f8-898d-4441f23b2245-kube-api-access-7mm4v\") pod \"redhat-operators-ftdpc\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:09:59 crc kubenswrapper[4744]: I0106 16:09:59.416578 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:10:00 crc kubenswrapper[4744]: I0106 16:10:00.002552 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ftdpc"] Jan 06 16:10:00 crc kubenswrapper[4744]: I0106 16:10:00.311097 4744 generic.go:334] "Generic (PLEG): container finished" podID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerID="3ff3170a5bd9724c36a449e64953897a717818d4ca82dbe06c1e22841deae541" exitCode=0 Jan 06 16:10:00 crc kubenswrapper[4744]: I0106 16:10:00.311201 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftdpc" event={"ID":"3d08e4f4-7126-49f8-898d-4441f23b2245","Type":"ContainerDied","Data":"3ff3170a5bd9724c36a449e64953897a717818d4ca82dbe06c1e22841deae541"} Jan 06 16:10:00 crc kubenswrapper[4744]: I0106 16:10:00.311228 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftdpc" event={"ID":"3d08e4f4-7126-49f8-898d-4441f23b2245","Type":"ContainerStarted","Data":"f8e3694a261365c5403038fdc844d7390bdd87c783d7956472081639db555ec3"} Jan 06 16:10:00 crc kubenswrapper[4744]: I0106 16:10:00.316453 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 16:10:01 crc kubenswrapper[4744]: I0106 16:10:01.325991 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftdpc" event={"ID":"3d08e4f4-7126-49f8-898d-4441f23b2245","Type":"ContainerStarted","Data":"b89ff524f0c7b98c56a750223ec0ea9cc1442b1492314c8a34108760f6969d3a"} Jan 06 16:10:05 crc kubenswrapper[4744]: I0106 16:10:05.388507 4744 generic.go:334] "Generic (PLEG): container finished" podID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerID="b89ff524f0c7b98c56a750223ec0ea9cc1442b1492314c8a34108760f6969d3a" exitCode=0 Jan 06 16:10:05 crc kubenswrapper[4744]: I0106 16:10:05.388612 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftdpc" event={"ID":"3d08e4f4-7126-49f8-898d-4441f23b2245","Type":"ContainerDied","Data":"b89ff524f0c7b98c56a750223ec0ea9cc1442b1492314c8a34108760f6969d3a"} Jan 06 16:10:06 crc kubenswrapper[4744]: I0106 16:10:06.402056 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftdpc" event={"ID":"3d08e4f4-7126-49f8-898d-4441f23b2245","Type":"ContainerStarted","Data":"b4a5345302d8e000d11bf66601b7b11953af39cc04aa93586a8b1d03b4f05df8"} Jan 06 16:10:06 crc kubenswrapper[4744]: I0106 16:10:06.424504 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ftdpc" podStartSLOduration=2.811104335 podStartE2EDuration="8.424481597s" podCreationTimestamp="2026-01-06 16:09:58 +0000 UTC" firstStartedPulling="2026-01-06 16:10:00.316003781 +0000 UTC m=+5596.943470139" lastFinishedPulling="2026-01-06 16:10:05.929381083 +0000 UTC m=+5602.556847401" observedRunningTime="2026-01-06 16:10:06.416806503 +0000 UTC m=+5603.044272821" watchObservedRunningTime="2026-01-06 16:10:06.424481597 +0000 UTC m=+5603.051947925" Jan 06 16:10:09 crc kubenswrapper[4744]: I0106 16:10:09.418089 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:10:09 crc kubenswrapper[4744]: I0106 16:10:09.418650 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:10:10 crc kubenswrapper[4744]: I0106 16:10:10.487014 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ftdpc" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="registry-server" probeResult="failure" output=< Jan 06 16:10:10 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 16:10:10 crc kubenswrapper[4744]: > Jan 06 16:10:20 crc kubenswrapper[4744]: I0106 16:10:20.485254 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ftdpc" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="registry-server" probeResult="failure" output=< Jan 06 16:10:20 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 16:10:20 crc kubenswrapper[4744]: > Jan 06 16:10:29 crc kubenswrapper[4744]: I0106 16:10:29.493861 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:10:29 crc kubenswrapper[4744]: I0106 16:10:29.563920 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:10:30 crc kubenswrapper[4744]: I0106 16:10:30.390782 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ftdpc"] Jan 06 16:10:30 crc kubenswrapper[4744]: I0106 16:10:30.740869 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ftdpc" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="registry-server" containerID="cri-o://b4a5345302d8e000d11bf66601b7b11953af39cc04aa93586a8b1d03b4f05df8" gracePeriod=2 Jan 06 16:10:31 crc kubenswrapper[4744]: I0106 16:10:31.758351 4744 generic.go:334] "Generic (PLEG): container finished" podID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerID="b4a5345302d8e000d11bf66601b7b11953af39cc04aa93586a8b1d03b4f05df8" exitCode=0 Jan 06 16:10:31 crc kubenswrapper[4744]: I0106 16:10:31.758498 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftdpc" event={"ID":"3d08e4f4-7126-49f8-898d-4441f23b2245","Type":"ContainerDied","Data":"b4a5345302d8e000d11bf66601b7b11953af39cc04aa93586a8b1d03b4f05df8"} Jan 06 16:10:31 crc kubenswrapper[4744]: I0106 16:10:31.991799 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.100266 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-utilities\") pod \"3d08e4f4-7126-49f8-898d-4441f23b2245\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.100561 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mm4v\" (UniqueName: \"kubernetes.io/projected/3d08e4f4-7126-49f8-898d-4441f23b2245-kube-api-access-7mm4v\") pod \"3d08e4f4-7126-49f8-898d-4441f23b2245\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.100596 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-catalog-content\") pod \"3d08e4f4-7126-49f8-898d-4441f23b2245\" (UID: \"3d08e4f4-7126-49f8-898d-4441f23b2245\") " Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.101265 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-utilities" (OuterVolumeSpecName: "utilities") pod "3d08e4f4-7126-49f8-898d-4441f23b2245" (UID: "3d08e4f4-7126-49f8-898d-4441f23b2245"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.107436 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d08e4f4-7126-49f8-898d-4441f23b2245-kube-api-access-7mm4v" (OuterVolumeSpecName: "kube-api-access-7mm4v") pod "3d08e4f4-7126-49f8-898d-4441f23b2245" (UID: "3d08e4f4-7126-49f8-898d-4441f23b2245"). InnerVolumeSpecName "kube-api-access-7mm4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.204076 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.204109 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mm4v\" (UniqueName: \"kubernetes.io/projected/3d08e4f4-7126-49f8-898d-4441f23b2245-kube-api-access-7mm4v\") on node \"crc\" DevicePath \"\"" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.207569 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d08e4f4-7126-49f8-898d-4441f23b2245" (UID: "3d08e4f4-7126-49f8-898d-4441f23b2245"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.305987 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d08e4f4-7126-49f8-898d-4441f23b2245-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.775221 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ftdpc" event={"ID":"3d08e4f4-7126-49f8-898d-4441f23b2245","Type":"ContainerDied","Data":"f8e3694a261365c5403038fdc844d7390bdd87c783d7956472081639db555ec3"} Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.775286 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ftdpc" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.775410 4744 scope.go:117] "RemoveContainer" containerID="b4a5345302d8e000d11bf66601b7b11953af39cc04aa93586a8b1d03b4f05df8" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.826308 4744 scope.go:117] "RemoveContainer" containerID="b89ff524f0c7b98c56a750223ec0ea9cc1442b1492314c8a34108760f6969d3a" Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.850713 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ftdpc"] Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.868248 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ftdpc"] Jan 06 16:10:32 crc kubenswrapper[4744]: I0106 16:10:32.908448 4744 scope.go:117] "RemoveContainer" containerID="3ff3170a5bd9724c36a449e64953897a717818d4ca82dbe06c1e22841deae541" Jan 06 16:10:33 crc kubenswrapper[4744]: I0106 16:10:33.730082 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" path="/var/lib/kubelet/pods/3d08e4f4-7126-49f8-898d-4441f23b2245/volumes" Jan 06 16:11:14 crc kubenswrapper[4744]: I0106 16:11:14.424388 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:11:14 crc kubenswrapper[4744]: I0106 16:11:14.425769 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:11:44 crc kubenswrapper[4744]: I0106 16:11:44.423364 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:11:44 crc kubenswrapper[4744]: I0106 16:11:44.424041 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:12:14 crc kubenswrapper[4744]: I0106 16:12:14.423375 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:12:14 crc kubenswrapper[4744]: I0106 16:12:14.423896 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:12:14 crc kubenswrapper[4744]: I0106 16:12:14.423948 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 16:12:14 crc kubenswrapper[4744]: I0106 16:12:14.424621 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 16:12:14 crc kubenswrapper[4744]: I0106 16:12:14.424685 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" gracePeriod=600 Jan 06 16:12:14 crc kubenswrapper[4744]: E0106 16:12:14.672333 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:12:15 crc kubenswrapper[4744]: I0106 16:12:15.144665 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" exitCode=0 Jan 06 16:12:15 crc kubenswrapper[4744]: I0106 16:12:15.144735 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb"} Jan 06 16:12:15 crc kubenswrapper[4744]: I0106 16:12:15.144788 4744 scope.go:117] "RemoveContainer" containerID="1fce60955246c02a8c5fa277cc49d89bc99389098884f65539257aa5968cb082" Jan 06 16:12:15 crc kubenswrapper[4744]: I0106 16:12:15.145629 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:12:15 crc kubenswrapper[4744]: E0106 16:12:15.146364 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:12:29 crc kubenswrapper[4744]: I0106 16:12:29.712179 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:12:29 crc kubenswrapper[4744]: E0106 16:12:29.713068 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.741486 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:12:43 crc kubenswrapper[4744]: E0106 16:12:43.744624 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.859876 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lcgh8"] Jan 06 16:12:43 crc kubenswrapper[4744]: E0106 16:12:43.860444 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="extract-utilities" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.860461 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="extract-utilities" Jan 06 16:12:43 crc kubenswrapper[4744]: E0106 16:12:43.860503 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="extract-content" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.860511 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="extract-content" Jan 06 16:12:43 crc kubenswrapper[4744]: E0106 16:12:43.860537 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="registry-server" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.860545 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="registry-server" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.860836 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d08e4f4-7126-49f8-898d-4441f23b2245" containerName="registry-server" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.862810 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.868358 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lcgh8"] Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.989113 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-catalog-content\") pod \"certified-operators-lcgh8\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.989270 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f92mm\" (UniqueName: \"kubernetes.io/projected/e59b9e13-85fc-4459-8fc8-c676678f3932-kube-api-access-f92mm\") pod \"certified-operators-lcgh8\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:43 crc kubenswrapper[4744]: I0106 16:12:43.989318 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-utilities\") pod \"certified-operators-lcgh8\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:44 crc kubenswrapper[4744]: I0106 16:12:44.091317 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f92mm\" (UniqueName: \"kubernetes.io/projected/e59b9e13-85fc-4459-8fc8-c676678f3932-kube-api-access-f92mm\") pod \"certified-operators-lcgh8\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:44 crc kubenswrapper[4744]: I0106 16:12:44.091417 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-utilities\") pod \"certified-operators-lcgh8\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:44 crc kubenswrapper[4744]: I0106 16:12:44.091518 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-catalog-content\") pod \"certified-operators-lcgh8\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:44 crc kubenswrapper[4744]: I0106 16:12:44.092033 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-catalog-content\") pod \"certified-operators-lcgh8\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:44 crc kubenswrapper[4744]: I0106 16:12:44.092355 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-utilities\") pod \"certified-operators-lcgh8\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:44 crc kubenswrapper[4744]: I0106 16:12:44.114312 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f92mm\" (UniqueName: \"kubernetes.io/projected/e59b9e13-85fc-4459-8fc8-c676678f3932-kube-api-access-f92mm\") pod \"certified-operators-lcgh8\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:44 crc kubenswrapper[4744]: I0106 16:12:44.211423 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:44 crc kubenswrapper[4744]: I0106 16:12:44.688398 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lcgh8"] Jan 06 16:12:45 crc kubenswrapper[4744]: I0106 16:12:45.533418 4744 generic.go:334] "Generic (PLEG): container finished" podID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerID="82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba" exitCode=0 Jan 06 16:12:45 crc kubenswrapper[4744]: I0106 16:12:45.533500 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcgh8" event={"ID":"e59b9e13-85fc-4459-8fc8-c676678f3932","Type":"ContainerDied","Data":"82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba"} Jan 06 16:12:45 crc kubenswrapper[4744]: I0106 16:12:45.533779 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcgh8" event={"ID":"e59b9e13-85fc-4459-8fc8-c676678f3932","Type":"ContainerStarted","Data":"84cce46568f31c49ab935fa1b21ccde7f81b4bb7e9e5f3c4675f89564b3c0127"} Jan 06 16:12:47 crc kubenswrapper[4744]: I0106 16:12:47.567905 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcgh8" event={"ID":"e59b9e13-85fc-4459-8fc8-c676678f3932","Type":"ContainerStarted","Data":"5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199"} Jan 06 16:12:48 crc kubenswrapper[4744]: I0106 16:12:48.583011 4744 generic.go:334] "Generic (PLEG): container finished" podID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerID="5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199" exitCode=0 Jan 06 16:12:48 crc kubenswrapper[4744]: I0106 16:12:48.583119 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcgh8" event={"ID":"e59b9e13-85fc-4459-8fc8-c676678f3932","Type":"ContainerDied","Data":"5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199"} Jan 06 16:12:49 crc kubenswrapper[4744]: I0106 16:12:49.599763 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcgh8" event={"ID":"e59b9e13-85fc-4459-8fc8-c676678f3932","Type":"ContainerStarted","Data":"0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c"} Jan 06 16:12:49 crc kubenswrapper[4744]: I0106 16:12:49.635517 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lcgh8" podStartSLOduration=2.949640804 podStartE2EDuration="6.635495116s" podCreationTimestamp="2026-01-06 16:12:43 +0000 UTC" firstStartedPulling="2026-01-06 16:12:45.537193102 +0000 UTC m=+5762.164659420" lastFinishedPulling="2026-01-06 16:12:49.223047384 +0000 UTC m=+5765.850513732" observedRunningTime="2026-01-06 16:12:49.62702245 +0000 UTC m=+5766.254488788" watchObservedRunningTime="2026-01-06 16:12:49.635495116 +0000 UTC m=+5766.262961444" Jan 06 16:12:54 crc kubenswrapper[4744]: I0106 16:12:54.211949 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:54 crc kubenswrapper[4744]: I0106 16:12:54.212695 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:54 crc kubenswrapper[4744]: I0106 16:12:54.294190 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:54 crc kubenswrapper[4744]: I0106 16:12:54.725875 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:54 crc kubenswrapper[4744]: I0106 16:12:54.787852 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lcgh8"] Jan 06 16:12:56 crc kubenswrapper[4744]: I0106 16:12:56.679310 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lcgh8" podUID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerName="registry-server" containerID="cri-o://0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c" gracePeriod=2 Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.637104 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.694008 4744 generic.go:334] "Generic (PLEG): container finished" podID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerID="0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c" exitCode=0 Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.694054 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcgh8" event={"ID":"e59b9e13-85fc-4459-8fc8-c676678f3932","Type":"ContainerDied","Data":"0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c"} Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.694084 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lcgh8" event={"ID":"e59b9e13-85fc-4459-8fc8-c676678f3932","Type":"ContainerDied","Data":"84cce46568f31c49ab935fa1b21ccde7f81b4bb7e9e5f3c4675f89564b3c0127"} Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.694101 4744 scope.go:117] "RemoveContainer" containerID="0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.694282 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lcgh8" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.717553 4744 scope.go:117] "RemoveContainer" containerID="5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.742768 4744 scope.go:117] "RemoveContainer" containerID="82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.765627 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f92mm\" (UniqueName: \"kubernetes.io/projected/e59b9e13-85fc-4459-8fc8-c676678f3932-kube-api-access-f92mm\") pod \"e59b9e13-85fc-4459-8fc8-c676678f3932\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.765748 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-utilities\") pod \"e59b9e13-85fc-4459-8fc8-c676678f3932\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.766042 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-catalog-content\") pod \"e59b9e13-85fc-4459-8fc8-c676678f3932\" (UID: \"e59b9e13-85fc-4459-8fc8-c676678f3932\") " Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.766940 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-utilities" (OuterVolumeSpecName: "utilities") pod "e59b9e13-85fc-4459-8fc8-c676678f3932" (UID: "e59b9e13-85fc-4459-8fc8-c676678f3932"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.774368 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e59b9e13-85fc-4459-8fc8-c676678f3932-kube-api-access-f92mm" (OuterVolumeSpecName: "kube-api-access-f92mm") pod "e59b9e13-85fc-4459-8fc8-c676678f3932" (UID: "e59b9e13-85fc-4459-8fc8-c676678f3932"). InnerVolumeSpecName "kube-api-access-f92mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.820509 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e59b9e13-85fc-4459-8fc8-c676678f3932" (UID: "e59b9e13-85fc-4459-8fc8-c676678f3932"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.838855 4744 scope.go:117] "RemoveContainer" containerID="0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c" Jan 06 16:12:57 crc kubenswrapper[4744]: E0106 16:12:57.839320 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c\": container with ID starting with 0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c not found: ID does not exist" containerID="0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.839370 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c"} err="failed to get container status \"0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c\": rpc error: code = NotFound desc = could not find container \"0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c\": container with ID starting with 0d32d36786c4ea386555c82981a0a056b89f2c05ed8f4f5265640444f510ff9c not found: ID does not exist" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.839401 4744 scope.go:117] "RemoveContainer" containerID="5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199" Jan 06 16:12:57 crc kubenswrapper[4744]: E0106 16:12:57.839708 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199\": container with ID starting with 5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199 not found: ID does not exist" containerID="5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.839740 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199"} err="failed to get container status \"5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199\": rpc error: code = NotFound desc = could not find container \"5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199\": container with ID starting with 5d886479a18b419331e91aa48784dd5c59b064a6591b3fd6c1bd06dbf8805199 not found: ID does not exist" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.839763 4744 scope.go:117] "RemoveContainer" containerID="82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba" Jan 06 16:12:57 crc kubenswrapper[4744]: E0106 16:12:57.840015 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba\": container with ID starting with 82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba not found: ID does not exist" containerID="82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.840055 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba"} err="failed to get container status \"82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba\": rpc error: code = NotFound desc = could not find container \"82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba\": container with ID starting with 82b4abbeb5964547248983543b3cfd67a22c6c0a96693d20a3e26d8d4e0625ba not found: ID does not exist" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.869426 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f92mm\" (UniqueName: \"kubernetes.io/projected/e59b9e13-85fc-4459-8fc8-c676678f3932-kube-api-access-f92mm\") on node \"crc\" DevicePath \"\"" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.869458 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:12:57 crc kubenswrapper[4744]: I0106 16:12:57.869467 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59b9e13-85fc-4459-8fc8-c676678f3932-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:12:58 crc kubenswrapper[4744]: I0106 16:12:58.033198 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lcgh8"] Jan 06 16:12:58 crc kubenswrapper[4744]: I0106 16:12:58.053618 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lcgh8"] Jan 06 16:12:58 crc kubenswrapper[4744]: I0106 16:12:58.712316 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:12:58 crc kubenswrapper[4744]: E0106 16:12:58.712841 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:12:59 crc kubenswrapper[4744]: I0106 16:12:59.728072 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e59b9e13-85fc-4459-8fc8-c676678f3932" path="/var/lib/kubelet/pods/e59b9e13-85fc-4459-8fc8-c676678f3932/volumes" Jan 06 16:13:13 crc kubenswrapper[4744]: I0106 16:13:13.711383 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:13:13 crc kubenswrapper[4744]: E0106 16:13:13.712645 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:13:27 crc kubenswrapper[4744]: I0106 16:13:27.718307 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:13:27 crc kubenswrapper[4744]: E0106 16:13:27.722600 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:13:39 crc kubenswrapper[4744]: I0106 16:13:39.711358 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:13:39 crc kubenswrapper[4744]: E0106 16:13:39.712218 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:13:51 crc kubenswrapper[4744]: I0106 16:13:51.712098 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:13:51 crc kubenswrapper[4744]: E0106 16:13:51.713242 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.788209 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xlx86"] Jan 06 16:14:05 crc kubenswrapper[4744]: E0106 16:14:05.789365 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerName="extract-content" Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.789388 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerName="extract-content" Jan 06 16:14:05 crc kubenswrapper[4744]: E0106 16:14:05.789426 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerName="extract-utilities" Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.789439 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerName="extract-utilities" Jan 06 16:14:05 crc kubenswrapper[4744]: E0106 16:14:05.789463 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerName="registry-server" Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.789472 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerName="registry-server" Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.789879 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="e59b9e13-85fc-4459-8fc8-c676678f3932" containerName="registry-server" Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.792322 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.818538 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlx86"] Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.900224 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-utilities\") pod \"redhat-marketplace-xlx86\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.901142 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltlv8\" (UniqueName: \"kubernetes.io/projected/8e916a99-b527-4f6d-8bff-9cd05e720513-kube-api-access-ltlv8\") pod \"redhat-marketplace-xlx86\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:05 crc kubenswrapper[4744]: I0106 16:14:05.901690 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-catalog-content\") pod \"redhat-marketplace-xlx86\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:06 crc kubenswrapper[4744]: I0106 16:14:06.004639 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltlv8\" (UniqueName: \"kubernetes.io/projected/8e916a99-b527-4f6d-8bff-9cd05e720513-kube-api-access-ltlv8\") pod \"redhat-marketplace-xlx86\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:06 crc kubenswrapper[4744]: I0106 16:14:06.004750 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-catalog-content\") pod \"redhat-marketplace-xlx86\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:06 crc kubenswrapper[4744]: I0106 16:14:06.004943 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-utilities\") pod \"redhat-marketplace-xlx86\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:06 crc kubenswrapper[4744]: I0106 16:14:06.005645 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-utilities\") pod \"redhat-marketplace-xlx86\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:06 crc kubenswrapper[4744]: I0106 16:14:06.006276 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-catalog-content\") pod \"redhat-marketplace-xlx86\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:06 crc kubenswrapper[4744]: I0106 16:14:06.038404 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltlv8\" (UniqueName: \"kubernetes.io/projected/8e916a99-b527-4f6d-8bff-9cd05e720513-kube-api-access-ltlv8\") pod \"redhat-marketplace-xlx86\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:06 crc kubenswrapper[4744]: I0106 16:14:06.126907 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:06 crc kubenswrapper[4744]: I0106 16:14:06.644051 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlx86"] Jan 06 16:14:06 crc kubenswrapper[4744]: W0106 16:14:06.663650 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e916a99_b527_4f6d_8bff_9cd05e720513.slice/crio-63c83de0ce3fc39b55d44f0fdf12fc8983f2d499e4389818864c71d9da4c4752 WatchSource:0}: Error finding container 63c83de0ce3fc39b55d44f0fdf12fc8983f2d499e4389818864c71d9da4c4752: Status 404 returned error can't find the container with id 63c83de0ce3fc39b55d44f0fdf12fc8983f2d499e4389818864c71d9da4c4752 Jan 06 16:14:06 crc kubenswrapper[4744]: I0106 16:14:06.711010 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:14:06 crc kubenswrapper[4744]: E0106 16:14:06.711455 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:14:07 crc kubenswrapper[4744]: E0106 16:14:07.508076 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.22:38228->38.102.83.22:45165: write tcp 38.102.83.22:38228->38.102.83.22:45165: write: connection reset by peer Jan 06 16:14:07 crc kubenswrapper[4744]: I0106 16:14:07.692347 4744 generic.go:334] "Generic (PLEG): container finished" podID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerID="acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078" exitCode=0 Jan 06 16:14:07 crc kubenswrapper[4744]: I0106 16:14:07.692395 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlx86" event={"ID":"8e916a99-b527-4f6d-8bff-9cd05e720513","Type":"ContainerDied","Data":"acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078"} Jan 06 16:14:07 crc kubenswrapper[4744]: I0106 16:14:07.692434 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlx86" event={"ID":"8e916a99-b527-4f6d-8bff-9cd05e720513","Type":"ContainerStarted","Data":"63c83de0ce3fc39b55d44f0fdf12fc8983f2d499e4389818864c71d9da4c4752"} Jan 06 16:14:09 crc kubenswrapper[4744]: I0106 16:14:09.734932 4744 generic.go:334] "Generic (PLEG): container finished" podID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerID="9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352" exitCode=0 Jan 06 16:14:09 crc kubenswrapper[4744]: I0106 16:14:09.735069 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlx86" event={"ID":"8e916a99-b527-4f6d-8bff-9cd05e720513","Type":"ContainerDied","Data":"9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352"} Jan 06 16:14:11 crc kubenswrapper[4744]: I0106 16:14:11.773243 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlx86" event={"ID":"8e916a99-b527-4f6d-8bff-9cd05e720513","Type":"ContainerStarted","Data":"ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41"} Jan 06 16:14:11 crc kubenswrapper[4744]: I0106 16:14:11.810478 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xlx86" podStartSLOduration=4.309301107 podStartE2EDuration="6.810449885s" podCreationTimestamp="2026-01-06 16:14:05 +0000 UTC" firstStartedPulling="2026-01-06 16:14:07.694314606 +0000 UTC m=+5844.321780914" lastFinishedPulling="2026-01-06 16:14:10.195463344 +0000 UTC m=+5846.822929692" observedRunningTime="2026-01-06 16:14:11.800446318 +0000 UTC m=+5848.427912676" watchObservedRunningTime="2026-01-06 16:14:11.810449885 +0000 UTC m=+5848.437916243" Jan 06 16:14:16 crc kubenswrapper[4744]: I0106 16:14:16.129337 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:16 crc kubenswrapper[4744]: I0106 16:14:16.130008 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:16 crc kubenswrapper[4744]: I0106 16:14:16.214533 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:16 crc kubenswrapper[4744]: I0106 16:14:16.927530 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:16 crc kubenswrapper[4744]: I0106 16:14:16.987633 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlx86"] Jan 06 16:14:18 crc kubenswrapper[4744]: I0106 16:14:18.711230 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:14:18 crc kubenswrapper[4744]: E0106 16:14:18.712071 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:14:18 crc kubenswrapper[4744]: I0106 16:14:18.876061 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xlx86" podUID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerName="registry-server" containerID="cri-o://ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41" gracePeriod=2 Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.449939 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.503119 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltlv8\" (UniqueName: \"kubernetes.io/projected/8e916a99-b527-4f6d-8bff-9cd05e720513-kube-api-access-ltlv8\") pod \"8e916a99-b527-4f6d-8bff-9cd05e720513\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.503799 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-catalog-content\") pod \"8e916a99-b527-4f6d-8bff-9cd05e720513\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.504056 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-utilities\") pod \"8e916a99-b527-4f6d-8bff-9cd05e720513\" (UID: \"8e916a99-b527-4f6d-8bff-9cd05e720513\") " Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.505470 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-utilities" (OuterVolumeSpecName: "utilities") pod "8e916a99-b527-4f6d-8bff-9cd05e720513" (UID: "8e916a99-b527-4f6d-8bff-9cd05e720513"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.513849 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e916a99-b527-4f6d-8bff-9cd05e720513-kube-api-access-ltlv8" (OuterVolumeSpecName: "kube-api-access-ltlv8") pod "8e916a99-b527-4f6d-8bff-9cd05e720513" (UID: "8e916a99-b527-4f6d-8bff-9cd05e720513"). InnerVolumeSpecName "kube-api-access-ltlv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.544827 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e916a99-b527-4f6d-8bff-9cd05e720513" (UID: "8e916a99-b527-4f6d-8bff-9cd05e720513"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.608050 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.608097 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e916a99-b527-4f6d-8bff-9cd05e720513-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.608111 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltlv8\" (UniqueName: \"kubernetes.io/projected/8e916a99-b527-4f6d-8bff-9cd05e720513-kube-api-access-ltlv8\") on node \"crc\" DevicePath \"\"" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.889855 4744 generic.go:334] "Generic (PLEG): container finished" podID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerID="ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41" exitCode=0 Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.889952 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlx86" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.889935 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlx86" event={"ID":"8e916a99-b527-4f6d-8bff-9cd05e720513","Type":"ContainerDied","Data":"ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41"} Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.890327 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlx86" event={"ID":"8e916a99-b527-4f6d-8bff-9cd05e720513","Type":"ContainerDied","Data":"63c83de0ce3fc39b55d44f0fdf12fc8983f2d499e4389818864c71d9da4c4752"} Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.890356 4744 scope.go:117] "RemoveContainer" containerID="ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.924757 4744 scope.go:117] "RemoveContainer" containerID="9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352" Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.937644 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlx86"] Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.964264 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlx86"] Jan 06 16:14:19 crc kubenswrapper[4744]: I0106 16:14:19.966476 4744 scope.go:117] "RemoveContainer" containerID="acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078" Jan 06 16:14:20 crc kubenswrapper[4744]: I0106 16:14:20.018372 4744 scope.go:117] "RemoveContainer" containerID="ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41" Jan 06 16:14:20 crc kubenswrapper[4744]: E0106 16:14:20.019236 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41\": container with ID starting with ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41 not found: ID does not exist" containerID="ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41" Jan 06 16:14:20 crc kubenswrapper[4744]: I0106 16:14:20.019334 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41"} err="failed to get container status \"ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41\": rpc error: code = NotFound desc = could not find container \"ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41\": container with ID starting with ea013e7f6084a0cedd56d9526b20d22b9dde023c08d26b47096317006c001d41 not found: ID does not exist" Jan 06 16:14:20 crc kubenswrapper[4744]: I0106 16:14:20.019406 4744 scope.go:117] "RemoveContainer" containerID="9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352" Jan 06 16:14:20 crc kubenswrapper[4744]: E0106 16:14:20.019836 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352\": container with ID starting with 9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352 not found: ID does not exist" containerID="9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352" Jan 06 16:14:20 crc kubenswrapper[4744]: I0106 16:14:20.019904 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352"} err="failed to get container status \"9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352\": rpc error: code = NotFound desc = could not find container \"9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352\": container with ID starting with 9c546e0264b145e3b30601bb5deb6b60faf461f484291caa1087fda468e84352 not found: ID does not exist" Jan 06 16:14:20 crc kubenswrapper[4744]: I0106 16:14:20.019922 4744 scope.go:117] "RemoveContainer" containerID="acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078" Jan 06 16:14:20 crc kubenswrapper[4744]: E0106 16:14:20.020376 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078\": container with ID starting with acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078 not found: ID does not exist" containerID="acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078" Jan 06 16:14:20 crc kubenswrapper[4744]: I0106 16:14:20.020442 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078"} err="failed to get container status \"acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078\": rpc error: code = NotFound desc = could not find container \"acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078\": container with ID starting with acf290b68ef7ec220861e6bd504fd807db56ff256d8ad782ee28c7bf0694a078 not found: ID does not exist" Jan 06 16:14:21 crc kubenswrapper[4744]: I0106 16:14:21.738046 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e916a99-b527-4f6d-8bff-9cd05e720513" path="/var/lib/kubelet/pods/8e916a99-b527-4f6d-8bff-9cd05e720513/volumes" Jan 06 16:14:29 crc kubenswrapper[4744]: I0106 16:14:29.712406 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:14:29 crc kubenswrapper[4744]: E0106 16:14:29.713913 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:14:40 crc kubenswrapper[4744]: I0106 16:14:40.711605 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:14:40 crc kubenswrapper[4744]: E0106 16:14:40.713046 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:14:52 crc kubenswrapper[4744]: I0106 16:14:52.711775 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:14:52 crc kubenswrapper[4744]: E0106 16:14:52.712665 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.167787 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8"] Jan 06 16:15:00 crc kubenswrapper[4744]: E0106 16:15:00.169175 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerName="extract-utilities" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.169194 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerName="extract-utilities" Jan 06 16:15:00 crc kubenswrapper[4744]: E0106 16:15:00.169214 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerName="registry-server" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.169244 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerName="registry-server" Jan 06 16:15:00 crc kubenswrapper[4744]: E0106 16:15:00.169272 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerName="extract-content" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.169283 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerName="extract-content" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.169656 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e916a99-b527-4f6d-8bff-9cd05e720513" containerName="registry-server" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.170649 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.175108 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.176399 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.197543 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8"] Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.273432 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2t7p\" (UniqueName: \"kubernetes.io/projected/3e647b26-53cf-4e64-9b38-65eeea502606-kube-api-access-t2t7p\") pod \"collect-profiles-29461935-jmmn8\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.273653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e647b26-53cf-4e64-9b38-65eeea502606-config-volume\") pod \"collect-profiles-29461935-jmmn8\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.273686 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e647b26-53cf-4e64-9b38-65eeea502606-secret-volume\") pod \"collect-profiles-29461935-jmmn8\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.377144 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e647b26-53cf-4e64-9b38-65eeea502606-config-volume\") pod \"collect-profiles-29461935-jmmn8\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.377258 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e647b26-53cf-4e64-9b38-65eeea502606-secret-volume\") pod \"collect-profiles-29461935-jmmn8\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.377568 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2t7p\" (UniqueName: \"kubernetes.io/projected/3e647b26-53cf-4e64-9b38-65eeea502606-kube-api-access-t2t7p\") pod \"collect-profiles-29461935-jmmn8\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.378118 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e647b26-53cf-4e64-9b38-65eeea502606-config-volume\") pod \"collect-profiles-29461935-jmmn8\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.383914 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e647b26-53cf-4e64-9b38-65eeea502606-secret-volume\") pod \"collect-profiles-29461935-jmmn8\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.397395 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2t7p\" (UniqueName: \"kubernetes.io/projected/3e647b26-53cf-4e64-9b38-65eeea502606-kube-api-access-t2t7p\") pod \"collect-profiles-29461935-jmmn8\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:00 crc kubenswrapper[4744]: I0106 16:15:00.497892 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:01 crc kubenswrapper[4744]: I0106 16:15:01.020902 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8"] Jan 06 16:15:01 crc kubenswrapper[4744]: I0106 16:15:01.465073 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" event={"ID":"3e647b26-53cf-4e64-9b38-65eeea502606","Type":"ContainerStarted","Data":"30d6c8fef64cc6d94f0f541bcd5bfaab03b5cec599559bec3f4df5d5ed1f8d1e"} Jan 06 16:15:01 crc kubenswrapper[4744]: I0106 16:15:01.465118 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" event={"ID":"3e647b26-53cf-4e64-9b38-65eeea502606","Type":"ContainerStarted","Data":"080877211a3ba07b09cfbb3c7a06442b4d2a30d86216e3dbd47784eb4b5230db"} Jan 06 16:15:02 crc kubenswrapper[4744]: I0106 16:15:02.475581 4744 generic.go:334] "Generic (PLEG): container finished" podID="3e647b26-53cf-4e64-9b38-65eeea502606" containerID="30d6c8fef64cc6d94f0f541bcd5bfaab03b5cec599559bec3f4df5d5ed1f8d1e" exitCode=0 Jan 06 16:15:02 crc kubenswrapper[4744]: I0106 16:15:02.475698 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" event={"ID":"3e647b26-53cf-4e64-9b38-65eeea502606","Type":"ContainerDied","Data":"30d6c8fef64cc6d94f0f541bcd5bfaab03b5cec599559bec3f4df5d5ed1f8d1e"} Jan 06 16:15:02 crc kubenswrapper[4744]: I0106 16:15:02.865410 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.046464 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2t7p\" (UniqueName: \"kubernetes.io/projected/3e647b26-53cf-4e64-9b38-65eeea502606-kube-api-access-t2t7p\") pod \"3e647b26-53cf-4e64-9b38-65eeea502606\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.046804 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e647b26-53cf-4e64-9b38-65eeea502606-config-volume\") pod \"3e647b26-53cf-4e64-9b38-65eeea502606\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.046918 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e647b26-53cf-4e64-9b38-65eeea502606-secret-volume\") pod \"3e647b26-53cf-4e64-9b38-65eeea502606\" (UID: \"3e647b26-53cf-4e64-9b38-65eeea502606\") " Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.047366 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e647b26-53cf-4e64-9b38-65eeea502606-config-volume" (OuterVolumeSpecName: "config-volume") pod "3e647b26-53cf-4e64-9b38-65eeea502606" (UID: "3e647b26-53cf-4e64-9b38-65eeea502606"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.047559 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3e647b26-53cf-4e64-9b38-65eeea502606-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.053149 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e647b26-53cf-4e64-9b38-65eeea502606-kube-api-access-t2t7p" (OuterVolumeSpecName: "kube-api-access-t2t7p") pod "3e647b26-53cf-4e64-9b38-65eeea502606" (UID: "3e647b26-53cf-4e64-9b38-65eeea502606"). InnerVolumeSpecName "kube-api-access-t2t7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.069580 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e647b26-53cf-4e64-9b38-65eeea502606-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3e647b26-53cf-4e64-9b38-65eeea502606" (UID: "3e647b26-53cf-4e64-9b38-65eeea502606"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.149517 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2t7p\" (UniqueName: \"kubernetes.io/projected/3e647b26-53cf-4e64-9b38-65eeea502606-kube-api-access-t2t7p\") on node \"crc\" DevicePath \"\"" Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.149552 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3e647b26-53cf-4e64-9b38-65eeea502606-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.490963 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" event={"ID":"3e647b26-53cf-4e64-9b38-65eeea502606","Type":"ContainerDied","Data":"080877211a3ba07b09cfbb3c7a06442b4d2a30d86216e3dbd47784eb4b5230db"} Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.491064 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="080877211a3ba07b09cfbb3c7a06442b4d2a30d86216e3dbd47784eb4b5230db" Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.491141 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461935-jmmn8" Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.961820 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k"] Jan 06 16:15:03 crc kubenswrapper[4744]: I0106 16:15:03.972560 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461890-vxh2k"] Jan 06 16:15:05 crc kubenswrapper[4744]: I0106 16:15:05.712210 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:15:05 crc kubenswrapper[4744]: E0106 16:15:05.713122 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:15:05 crc kubenswrapper[4744]: I0106 16:15:05.735764 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c7b5506-b57e-4203-85b7-0451d584b652" path="/var/lib/kubelet/pods/4c7b5506-b57e-4203-85b7-0451d584b652/volumes" Jan 06 16:15:14 crc kubenswrapper[4744]: I0106 16:15:14.412789 4744 scope.go:117] "RemoveContainer" containerID="2dede9b56059ba8afe86ff7853489d5dff6ffc9f1f0816a5e88b3e156436a3c4" Jan 06 16:15:19 crc kubenswrapper[4744]: I0106 16:15:19.713282 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:15:19 crc kubenswrapper[4744]: E0106 16:15:19.714547 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:15:34 crc kubenswrapper[4744]: I0106 16:15:34.712269 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:15:34 crc kubenswrapper[4744]: E0106 16:15:34.713094 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:15:45 crc kubenswrapper[4744]: I0106 16:15:45.712214 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:15:45 crc kubenswrapper[4744]: E0106 16:15:45.713460 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:15:57 crc kubenswrapper[4744]: I0106 16:15:57.711284 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:15:57 crc kubenswrapper[4744]: E0106 16:15:57.712126 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:16:08 crc kubenswrapper[4744]: I0106 16:16:08.711229 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:16:08 crc kubenswrapper[4744]: E0106 16:16:08.712202 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:16:10 crc kubenswrapper[4744]: I0106 16:16:10.758510 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="b2b4c15b-189d-41b2-b121-fdc25beb18a4" containerName="galera" probeResult="failure" output="command timed out" Jan 06 16:16:10 crc kubenswrapper[4744]: I0106 16:16:10.768538 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="b2b4c15b-189d-41b2-b121-fdc25beb18a4" containerName="galera" probeResult="failure" output="command timed out" Jan 06 16:16:21 crc kubenswrapper[4744]: I0106 16:16:21.711949 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:16:21 crc kubenswrapper[4744]: E0106 16:16:21.713018 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:16:32 crc kubenswrapper[4744]: I0106 16:16:32.712749 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:16:32 crc kubenswrapper[4744]: E0106 16:16:32.713632 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:16:45 crc kubenswrapper[4744]: I0106 16:16:45.711868 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:16:45 crc kubenswrapper[4744]: E0106 16:16:45.712535 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.711939 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:16:59 crc kubenswrapper[4744]: E0106 16:16:59.713298 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.775578 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bljm5"] Jan 06 16:16:59 crc kubenswrapper[4744]: E0106 16:16:59.776268 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e647b26-53cf-4e64-9b38-65eeea502606" containerName="collect-profiles" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.776293 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e647b26-53cf-4e64-9b38-65eeea502606" containerName="collect-profiles" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.776646 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e647b26-53cf-4e64-9b38-65eeea502606" containerName="collect-profiles" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.778748 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.792926 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bljm5"] Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.876140 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-utilities\") pod \"community-operators-bljm5\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.876454 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-catalog-content\") pod \"community-operators-bljm5\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.876713 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr9d8\" (UniqueName: \"kubernetes.io/projected/7ac287ae-48e3-4e50-9a5f-97f2965d296c-kube-api-access-sr9d8\") pod \"community-operators-bljm5\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.979499 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr9d8\" (UniqueName: \"kubernetes.io/projected/7ac287ae-48e3-4e50-9a5f-97f2965d296c-kube-api-access-sr9d8\") pod \"community-operators-bljm5\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.979661 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-utilities\") pod \"community-operators-bljm5\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.979718 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-catalog-content\") pod \"community-operators-bljm5\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.980302 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-utilities\") pod \"community-operators-bljm5\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:16:59 crc kubenswrapper[4744]: I0106 16:16:59.980506 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-catalog-content\") pod \"community-operators-bljm5\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:17:00 crc kubenswrapper[4744]: I0106 16:17:00.358939 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr9d8\" (UniqueName: \"kubernetes.io/projected/7ac287ae-48e3-4e50-9a5f-97f2965d296c-kube-api-access-sr9d8\") pod \"community-operators-bljm5\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:17:00 crc kubenswrapper[4744]: I0106 16:17:00.426378 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:17:01 crc kubenswrapper[4744]: I0106 16:17:01.003478 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bljm5"] Jan 06 16:17:01 crc kubenswrapper[4744]: I0106 16:17:01.066891 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bljm5" event={"ID":"7ac287ae-48e3-4e50-9a5f-97f2965d296c","Type":"ContainerStarted","Data":"7e1d1a2e3a025a5076b1403ad68d5335a81ff0adc026b38e799087394c9cc736"} Jan 06 16:17:02 crc kubenswrapper[4744]: I0106 16:17:02.119632 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerID="e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a" exitCode=0 Jan 06 16:17:02 crc kubenswrapper[4744]: I0106 16:17:02.120362 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bljm5" event={"ID":"7ac287ae-48e3-4e50-9a5f-97f2965d296c","Type":"ContainerDied","Data":"e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a"} Jan 06 16:17:02 crc kubenswrapper[4744]: I0106 16:17:02.123347 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 16:17:04 crc kubenswrapper[4744]: E0106 16:17:04.078862 4744 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.22:38456->38.102.83.22:45165: write tcp 38.102.83.22:38456->38.102.83.22:45165: write: broken pipe Jan 06 16:17:04 crc kubenswrapper[4744]: I0106 16:17:04.147724 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bljm5" event={"ID":"7ac287ae-48e3-4e50-9a5f-97f2965d296c","Type":"ContainerStarted","Data":"1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8"} Jan 06 16:17:06 crc kubenswrapper[4744]: I0106 16:17:06.173231 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerID="1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8" exitCode=0 Jan 06 16:17:06 crc kubenswrapper[4744]: I0106 16:17:06.173311 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bljm5" event={"ID":"7ac287ae-48e3-4e50-9a5f-97f2965d296c","Type":"ContainerDied","Data":"1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8"} Jan 06 16:17:07 crc kubenswrapper[4744]: I0106 16:17:07.190018 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bljm5" event={"ID":"7ac287ae-48e3-4e50-9a5f-97f2965d296c","Type":"ContainerStarted","Data":"9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c"} Jan 06 16:17:07 crc kubenswrapper[4744]: I0106 16:17:07.215365 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bljm5" podStartSLOduration=3.713101301 podStartE2EDuration="8.21534296s" podCreationTimestamp="2026-01-06 16:16:59 +0000 UTC" firstStartedPulling="2026-01-06 16:17:02.123110328 +0000 UTC m=+6018.750576646" lastFinishedPulling="2026-01-06 16:17:06.625351977 +0000 UTC m=+6023.252818305" observedRunningTime="2026-01-06 16:17:07.210588484 +0000 UTC m=+6023.838054822" watchObservedRunningTime="2026-01-06 16:17:07.21534296 +0000 UTC m=+6023.842809278" Jan 06 16:17:10 crc kubenswrapper[4744]: I0106 16:17:10.427338 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:17:10 crc kubenswrapper[4744]: I0106 16:17:10.428127 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:17:10 crc kubenswrapper[4744]: I0106 16:17:10.517279 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:17:11 crc kubenswrapper[4744]: I0106 16:17:11.293093 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:17:11 crc kubenswrapper[4744]: I0106 16:17:11.343381 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bljm5"] Jan 06 16:17:13 crc kubenswrapper[4744]: I0106 16:17:13.267989 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bljm5" podUID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerName="registry-server" containerID="cri-o://9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c" gracePeriod=2 Jan 06 16:17:13 crc kubenswrapper[4744]: I0106 16:17:13.857695 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:17:13 crc kubenswrapper[4744]: I0106 16:17:13.971445 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr9d8\" (UniqueName: \"kubernetes.io/projected/7ac287ae-48e3-4e50-9a5f-97f2965d296c-kube-api-access-sr9d8\") pod \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " Jan 06 16:17:13 crc kubenswrapper[4744]: I0106 16:17:13.971635 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-utilities\") pod \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " Jan 06 16:17:13 crc kubenswrapper[4744]: I0106 16:17:13.971737 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-catalog-content\") pod \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\" (UID: \"7ac287ae-48e3-4e50-9a5f-97f2965d296c\") " Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.224441 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-utilities" (OuterVolumeSpecName: "utilities") pod "7ac287ae-48e3-4e50-9a5f-97f2965d296c" (UID: "7ac287ae-48e3-4e50-9a5f-97f2965d296c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.230115 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac287ae-48e3-4e50-9a5f-97f2965d296c-kube-api-access-sr9d8" (OuterVolumeSpecName: "kube-api-access-sr9d8") pod "7ac287ae-48e3-4e50-9a5f-97f2965d296c" (UID: "7ac287ae-48e3-4e50-9a5f-97f2965d296c"). InnerVolumeSpecName "kube-api-access-sr9d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.281655 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr9d8\" (UniqueName: \"kubernetes.io/projected/7ac287ae-48e3-4e50-9a5f-97f2965d296c-kube-api-access-sr9d8\") on node \"crc\" DevicePath \"\"" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.281702 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.290126 4744 generic.go:334] "Generic (PLEG): container finished" podID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerID="9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c" exitCode=0 Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.290201 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bljm5" event={"ID":"7ac287ae-48e3-4e50-9a5f-97f2965d296c","Type":"ContainerDied","Data":"9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c"} Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.290232 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bljm5" event={"ID":"7ac287ae-48e3-4e50-9a5f-97f2965d296c","Type":"ContainerDied","Data":"7e1d1a2e3a025a5076b1403ad68d5335a81ff0adc026b38e799087394c9cc736"} Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.290252 4744 scope.go:117] "RemoveContainer" containerID="9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.290466 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bljm5" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.300384 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ac287ae-48e3-4e50-9a5f-97f2965d296c" (UID: "7ac287ae-48e3-4e50-9a5f-97f2965d296c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.344663 4744 scope.go:117] "RemoveContainer" containerID="1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.377908 4744 scope.go:117] "RemoveContainer" containerID="e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.383762 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ac287ae-48e3-4e50-9a5f-97f2965d296c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.416635 4744 scope.go:117] "RemoveContainer" containerID="9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c" Jan 06 16:17:14 crc kubenswrapper[4744]: E0106 16:17:14.418855 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c\": container with ID starting with 9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c not found: ID does not exist" containerID="9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.418910 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c"} err="failed to get container status \"9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c\": rpc error: code = NotFound desc = could not find container \"9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c\": container with ID starting with 9323cea7de8555b516919b8654d7304bb93cb47b48df8077cb09a64225d9298c not found: ID does not exist" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.418941 4744 scope.go:117] "RemoveContainer" containerID="1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8" Jan 06 16:17:14 crc kubenswrapper[4744]: E0106 16:17:14.422156 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8\": container with ID starting with 1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8 not found: ID does not exist" containerID="1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.422211 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8"} err="failed to get container status \"1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8\": rpc error: code = NotFound desc = could not find container \"1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8\": container with ID starting with 1bf79521b25d3e0d033e1b423169231318eb4c2ec1b384055062823cfff4b0a8 not found: ID does not exist" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.422228 4744 scope.go:117] "RemoveContainer" containerID="e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a" Jan 06 16:17:14 crc kubenswrapper[4744]: E0106 16:17:14.422619 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a\": container with ID starting with e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a not found: ID does not exist" containerID="e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.422655 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a"} err="failed to get container status \"e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a\": rpc error: code = NotFound desc = could not find container \"e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a\": container with ID starting with e935e44bda180005f1ff5011c5129d012a03613199599339c9733d0700c90b0a not found: ID does not exist" Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.622307 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bljm5"] Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.630995 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bljm5"] Jan 06 16:17:14 crc kubenswrapper[4744]: I0106 16:17:14.712511 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:17:15 crc kubenswrapper[4744]: I0106 16:17:15.308880 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"221b2e5281ccbb5da1e0a7b76c3b63ae8d2ef757f93ecb10652f51c87dfd189b"} Jan 06 16:17:15 crc kubenswrapper[4744]: I0106 16:17:15.724860 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" path="/var/lib/kubelet/pods/7ac287ae-48e3-4e50-9a5f-97f2965d296c/volumes" Jan 06 16:17:49 crc kubenswrapper[4744]: E0106 16:17:49.808569 4744 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.22:42948->38.102.83.22:45165: read tcp 38.102.83.22:42948->38.102.83.22:45165: read: connection reset by peer Jan 06 16:18:13 crc kubenswrapper[4744]: I0106 16:18:13.841386 4744 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-lmtg5 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.75:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 16:18:13 crc kubenswrapper[4744]: I0106 16:18:13.841768 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" podUID="c9626172-6c81-40bd-91a7-48e6790e9f7f" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.75:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 06 16:18:13 crc kubenswrapper[4744]: I0106 16:18:13.934744 4744 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-lmtg5 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.75:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 06 16:18:13 crc kubenswrapper[4744]: I0106 16:18:13.934822 4744 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-lmtg5" podUID="c9626172-6c81-40bd-91a7-48e6790e9f7f" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.75:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 06 16:19:13 crc kubenswrapper[4744]: I0106 16:19:13.312012 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-9l4fk" podUID="2bf34d55-4b6d-4716-aca0-026cdd02aabe" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 06 16:19:14 crc kubenswrapper[4744]: I0106 16:19:14.423747 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:19:14 crc kubenswrapper[4744]: I0106 16:19:14.424107 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:19:44 crc kubenswrapper[4744]: I0106 16:19:44.423862 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:19:44 crc kubenswrapper[4744]: I0106 16:19:44.424745 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.386547 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kg8vm"] Jan 06 16:20:01 crc kubenswrapper[4744]: E0106 16:20:01.390544 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerName="extract-content" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.390571 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerName="extract-content" Jan 06 16:20:01 crc kubenswrapper[4744]: E0106 16:20:01.390611 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerName="registry-server" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.390618 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerName="registry-server" Jan 06 16:20:01 crc kubenswrapper[4744]: E0106 16:20:01.390637 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerName="extract-utilities" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.390643 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerName="extract-utilities" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.390869 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac287ae-48e3-4e50-9a5f-97f2965d296c" containerName="registry-server" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.392515 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.420952 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kg8vm"] Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.479614 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qwrd\" (UniqueName: \"kubernetes.io/projected/023441d0-3240-4b95-9355-58080a7ccbd9-kube-api-access-7qwrd\") pod \"redhat-operators-kg8vm\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.479666 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-catalog-content\") pod \"redhat-operators-kg8vm\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.479763 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-utilities\") pod \"redhat-operators-kg8vm\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.581743 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-utilities\") pod \"redhat-operators-kg8vm\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.581962 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qwrd\" (UniqueName: \"kubernetes.io/projected/023441d0-3240-4b95-9355-58080a7ccbd9-kube-api-access-7qwrd\") pod \"redhat-operators-kg8vm\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.581992 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-catalog-content\") pod \"redhat-operators-kg8vm\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.582383 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-utilities\") pod \"redhat-operators-kg8vm\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.582390 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-catalog-content\") pod \"redhat-operators-kg8vm\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.602120 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qwrd\" (UniqueName: \"kubernetes.io/projected/023441d0-3240-4b95-9355-58080a7ccbd9-kube-api-access-7qwrd\") pod \"redhat-operators-kg8vm\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:01 crc kubenswrapper[4744]: I0106 16:20:01.720311 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:02 crc kubenswrapper[4744]: I0106 16:20:02.234333 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kg8vm"] Jan 06 16:20:02 crc kubenswrapper[4744]: W0106 16:20:02.234873 4744 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod023441d0_3240_4b95_9355_58080a7ccbd9.slice/crio-21bd907a47b7fe79e74f1679e4927465c23160f67f9e1318f43efc76164c472f WatchSource:0}: Error finding container 21bd907a47b7fe79e74f1679e4927465c23160f67f9e1318f43efc76164c472f: Status 404 returned error can't find the container with id 21bd907a47b7fe79e74f1679e4927465c23160f67f9e1318f43efc76164c472f Jan 06 16:20:02 crc kubenswrapper[4744]: I0106 16:20:02.312073 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg8vm" event={"ID":"023441d0-3240-4b95-9355-58080a7ccbd9","Type":"ContainerStarted","Data":"21bd907a47b7fe79e74f1679e4927465c23160f67f9e1318f43efc76164c472f"} Jan 06 16:20:03 crc kubenswrapper[4744]: I0106 16:20:03.324290 4744 generic.go:334] "Generic (PLEG): container finished" podID="023441d0-3240-4b95-9355-58080a7ccbd9" containerID="e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a" exitCode=0 Jan 06 16:20:03 crc kubenswrapper[4744]: I0106 16:20:03.325526 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg8vm" event={"ID":"023441d0-3240-4b95-9355-58080a7ccbd9","Type":"ContainerDied","Data":"e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a"} Jan 06 16:20:05 crc kubenswrapper[4744]: I0106 16:20:05.349573 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg8vm" event={"ID":"023441d0-3240-4b95-9355-58080a7ccbd9","Type":"ContainerStarted","Data":"b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab"} Jan 06 16:20:08 crc kubenswrapper[4744]: I0106 16:20:08.402736 4744 generic.go:334] "Generic (PLEG): container finished" podID="023441d0-3240-4b95-9355-58080a7ccbd9" containerID="b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab" exitCode=0 Jan 06 16:20:08 crc kubenswrapper[4744]: I0106 16:20:08.402818 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg8vm" event={"ID":"023441d0-3240-4b95-9355-58080a7ccbd9","Type":"ContainerDied","Data":"b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab"} Jan 06 16:20:09 crc kubenswrapper[4744]: I0106 16:20:09.424611 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg8vm" event={"ID":"023441d0-3240-4b95-9355-58080a7ccbd9","Type":"ContainerStarted","Data":"65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3"} Jan 06 16:20:09 crc kubenswrapper[4744]: I0106 16:20:09.451320 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kg8vm" podStartSLOduration=2.757255889 podStartE2EDuration="8.45130081s" podCreationTimestamp="2026-01-06 16:20:01 +0000 UTC" firstStartedPulling="2026-01-06 16:20:03.328069911 +0000 UTC m=+6199.955536239" lastFinishedPulling="2026-01-06 16:20:09.022114832 +0000 UTC m=+6205.649581160" observedRunningTime="2026-01-06 16:20:09.444827448 +0000 UTC m=+6206.072293766" watchObservedRunningTime="2026-01-06 16:20:09.45130081 +0000 UTC m=+6206.078767118" Jan 06 16:20:11 crc kubenswrapper[4744]: I0106 16:20:11.728019 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:11 crc kubenswrapper[4744]: I0106 16:20:11.728792 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:12 crc kubenswrapper[4744]: I0106 16:20:12.809078 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kg8vm" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" containerName="registry-server" probeResult="failure" output=< Jan 06 16:20:12 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 16:20:12 crc kubenswrapper[4744]: > Jan 06 16:20:14 crc kubenswrapper[4744]: I0106 16:20:14.424488 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:20:14 crc kubenswrapper[4744]: I0106 16:20:14.424782 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:20:14 crc kubenswrapper[4744]: I0106 16:20:14.424834 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 16:20:14 crc kubenswrapper[4744]: I0106 16:20:14.425778 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"221b2e5281ccbb5da1e0a7b76c3b63ae8d2ef757f93ecb10652f51c87dfd189b"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 16:20:14 crc kubenswrapper[4744]: I0106 16:20:14.425849 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://221b2e5281ccbb5da1e0a7b76c3b63ae8d2ef757f93ecb10652f51c87dfd189b" gracePeriod=600 Jan 06 16:20:15 crc kubenswrapper[4744]: I0106 16:20:15.508590 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="221b2e5281ccbb5da1e0a7b76c3b63ae8d2ef757f93ecb10652f51c87dfd189b" exitCode=0 Jan 06 16:20:15 crc kubenswrapper[4744]: I0106 16:20:15.508647 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"221b2e5281ccbb5da1e0a7b76c3b63ae8d2ef757f93ecb10652f51c87dfd189b"} Jan 06 16:20:15 crc kubenswrapper[4744]: I0106 16:20:15.509119 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f"} Jan 06 16:20:15 crc kubenswrapper[4744]: I0106 16:20:15.509153 4744 scope.go:117] "RemoveContainer" containerID="812facc68cf42006cde73889df9ccda0b8e692c4845929527a82ee75317d61bb" Jan 06 16:20:21 crc kubenswrapper[4744]: I0106 16:20:21.794371 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:21 crc kubenswrapper[4744]: I0106 16:20:21.865059 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:22 crc kubenswrapper[4744]: I0106 16:20:22.043750 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kg8vm"] Jan 06 16:20:23 crc kubenswrapper[4744]: I0106 16:20:23.606875 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kg8vm" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" containerName="registry-server" containerID="cri-o://65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3" gracePeriod=2 Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.208343 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.336373 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-catalog-content\") pod \"023441d0-3240-4b95-9355-58080a7ccbd9\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.336494 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-utilities\") pod \"023441d0-3240-4b95-9355-58080a7ccbd9\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.336645 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qwrd\" (UniqueName: \"kubernetes.io/projected/023441d0-3240-4b95-9355-58080a7ccbd9-kube-api-access-7qwrd\") pod \"023441d0-3240-4b95-9355-58080a7ccbd9\" (UID: \"023441d0-3240-4b95-9355-58080a7ccbd9\") " Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.337306 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-utilities" (OuterVolumeSpecName: "utilities") pod "023441d0-3240-4b95-9355-58080a7ccbd9" (UID: "023441d0-3240-4b95-9355-58080a7ccbd9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.338562 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.343504 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/023441d0-3240-4b95-9355-58080a7ccbd9-kube-api-access-7qwrd" (OuterVolumeSpecName: "kube-api-access-7qwrd") pod "023441d0-3240-4b95-9355-58080a7ccbd9" (UID: "023441d0-3240-4b95-9355-58080a7ccbd9"). InnerVolumeSpecName "kube-api-access-7qwrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.440829 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qwrd\" (UniqueName: \"kubernetes.io/projected/023441d0-3240-4b95-9355-58080a7ccbd9-kube-api-access-7qwrd\") on node \"crc\" DevicePath \"\"" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.457992 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "023441d0-3240-4b95-9355-58080a7ccbd9" (UID: "023441d0-3240-4b95-9355-58080a7ccbd9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.543127 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/023441d0-3240-4b95-9355-58080a7ccbd9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.621129 4744 generic.go:334] "Generic (PLEG): container finished" podID="023441d0-3240-4b95-9355-58080a7ccbd9" containerID="65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3" exitCode=0 Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.621189 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg8vm" event={"ID":"023441d0-3240-4b95-9355-58080a7ccbd9","Type":"ContainerDied","Data":"65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3"} Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.621218 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kg8vm" event={"ID":"023441d0-3240-4b95-9355-58080a7ccbd9","Type":"ContainerDied","Data":"21bd907a47b7fe79e74f1679e4927465c23160f67f9e1318f43efc76164c472f"} Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.621234 4744 scope.go:117] "RemoveContainer" containerID="65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.621413 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kg8vm" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.666758 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kg8vm"] Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.669920 4744 scope.go:117] "RemoveContainer" containerID="b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.679638 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kg8vm"] Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.712334 4744 scope.go:117] "RemoveContainer" containerID="e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.784412 4744 scope.go:117] "RemoveContainer" containerID="65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3" Jan 06 16:20:24 crc kubenswrapper[4744]: E0106 16:20:24.784988 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3\": container with ID starting with 65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3 not found: ID does not exist" containerID="65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.785031 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3"} err="failed to get container status \"65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3\": rpc error: code = NotFound desc = could not find container \"65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3\": container with ID starting with 65cde3d19b37b27c2c31e0a0bd46a1e5732c8dcecafca40f69e27545d0250cb3 not found: ID does not exist" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.785059 4744 scope.go:117] "RemoveContainer" containerID="b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab" Jan 06 16:20:24 crc kubenswrapper[4744]: E0106 16:20:24.785372 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab\": container with ID starting with b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab not found: ID does not exist" containerID="b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.785402 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab"} err="failed to get container status \"b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab\": rpc error: code = NotFound desc = could not find container \"b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab\": container with ID starting with b30c38eb70dba452b307f2c5107c4bf06d0d0916520c0320fffc50d537846aab not found: ID does not exist" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.785423 4744 scope.go:117] "RemoveContainer" containerID="e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a" Jan 06 16:20:24 crc kubenswrapper[4744]: E0106 16:20:24.785687 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a\": container with ID starting with e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a not found: ID does not exist" containerID="e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a" Jan 06 16:20:24 crc kubenswrapper[4744]: I0106 16:20:24.785712 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a"} err="failed to get container status \"e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a\": rpc error: code = NotFound desc = could not find container \"e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a\": container with ID starting with e65a03eb41a5ac55304f40d011db325c6ba587ae76ad00613ee5c4f787eb093a not found: ID does not exist" Jan 06 16:20:25 crc kubenswrapper[4744]: I0106 16:20:25.728008 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" path="/var/lib/kubelet/pods/023441d0-3240-4b95-9355-58080a7ccbd9/volumes" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.664687 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8vxws/must-gather-vqjqk"] Jan 06 16:20:31 crc kubenswrapper[4744]: E0106 16:20:31.665425 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" containerName="extract-content" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.665437 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" containerName="extract-content" Jan 06 16:20:31 crc kubenswrapper[4744]: E0106 16:20:31.665446 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" containerName="registry-server" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.665452 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" containerName="registry-server" Jan 06 16:20:31 crc kubenswrapper[4744]: E0106 16:20:31.665468 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" containerName="extract-utilities" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.665474 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" containerName="extract-utilities" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.665671 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="023441d0-3240-4b95-9355-58080a7ccbd9" containerName="registry-server" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.667116 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.669746 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8vxws"/"kube-root-ca.crt" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.669793 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8vxws"/"openshift-service-ca.crt" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.670020 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8vxws"/"default-dockercfg-n5rxc" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.691155 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8vxws/must-gather-vqjqk"] Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.774660 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d643fddf-ee4c-4226-814c-4acc9a13eb91-must-gather-output\") pod \"must-gather-vqjqk\" (UID: \"d643fddf-ee4c-4226-814c-4acc9a13eb91\") " pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.774709 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqjqv\" (UniqueName: \"kubernetes.io/projected/d643fddf-ee4c-4226-814c-4acc9a13eb91-kube-api-access-lqjqv\") pod \"must-gather-vqjqk\" (UID: \"d643fddf-ee4c-4226-814c-4acc9a13eb91\") " pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.877418 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d643fddf-ee4c-4226-814c-4acc9a13eb91-must-gather-output\") pod \"must-gather-vqjqk\" (UID: \"d643fddf-ee4c-4226-814c-4acc9a13eb91\") " pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.877702 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqjqv\" (UniqueName: \"kubernetes.io/projected/d643fddf-ee4c-4226-814c-4acc9a13eb91-kube-api-access-lqjqv\") pod \"must-gather-vqjqk\" (UID: \"d643fddf-ee4c-4226-814c-4acc9a13eb91\") " pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.877897 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d643fddf-ee4c-4226-814c-4acc9a13eb91-must-gather-output\") pod \"must-gather-vqjqk\" (UID: \"d643fddf-ee4c-4226-814c-4acc9a13eb91\") " pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.894682 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqjqv\" (UniqueName: \"kubernetes.io/projected/d643fddf-ee4c-4226-814c-4acc9a13eb91-kube-api-access-lqjqv\") pod \"must-gather-vqjqk\" (UID: \"d643fddf-ee4c-4226-814c-4acc9a13eb91\") " pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:20:31 crc kubenswrapper[4744]: I0106 16:20:31.983999 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:20:32 crc kubenswrapper[4744]: I0106 16:20:32.514294 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8vxws/must-gather-vqjqk"] Jan 06 16:20:32 crc kubenswrapper[4744]: I0106 16:20:32.756845 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/must-gather-vqjqk" event={"ID":"d643fddf-ee4c-4226-814c-4acc9a13eb91","Type":"ContainerStarted","Data":"a752260129c15d2f37056df76b4f131593ed1835367d6273380ac41c629e6da5"} Jan 06 16:20:41 crc kubenswrapper[4744]: I0106 16:20:41.899631 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/must-gather-vqjqk" event={"ID":"d643fddf-ee4c-4226-814c-4acc9a13eb91","Type":"ContainerStarted","Data":"4576b7e4ded6902d80bb24939deab06f7cef9aa30bf17f8bbe287a3819e73fe9"} Jan 06 16:20:41 crc kubenswrapper[4744]: I0106 16:20:41.900474 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/must-gather-vqjqk" event={"ID":"d643fddf-ee4c-4226-814c-4acc9a13eb91","Type":"ContainerStarted","Data":"bacda9e92884680457b14ff3b2d1ebdaea4498c05385abf6b2b2c9ef58a12109"} Jan 06 16:20:41 crc kubenswrapper[4744]: I0106 16:20:41.947652 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8vxws/must-gather-vqjqk" podStartSLOduration=3.407871656 podStartE2EDuration="10.947619647s" podCreationTimestamp="2026-01-06 16:20:31 +0000 UTC" firstStartedPulling="2026-01-06 16:20:32.518390029 +0000 UTC m=+6229.145856347" lastFinishedPulling="2026-01-06 16:20:40.05813802 +0000 UTC m=+6236.685604338" observedRunningTime="2026-01-06 16:20:41.925722244 +0000 UTC m=+6238.553188602" watchObservedRunningTime="2026-01-06 16:20:41.947619647 +0000 UTC m=+6238.575086005" Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.234150 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8vxws/crc-debug-g9mq7"] Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.236443 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.342733 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gndq8\" (UniqueName: \"kubernetes.io/projected/603a2d8e-2e64-4d45-9449-3e180dd9c340-kube-api-access-gndq8\") pod \"crc-debug-g9mq7\" (UID: \"603a2d8e-2e64-4d45-9449-3e180dd9c340\") " pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.342775 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603a2d8e-2e64-4d45-9449-3e180dd9c340-host\") pod \"crc-debug-g9mq7\" (UID: \"603a2d8e-2e64-4d45-9449-3e180dd9c340\") " pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.445500 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gndq8\" (UniqueName: \"kubernetes.io/projected/603a2d8e-2e64-4d45-9449-3e180dd9c340-kube-api-access-gndq8\") pod \"crc-debug-g9mq7\" (UID: \"603a2d8e-2e64-4d45-9449-3e180dd9c340\") " pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.445570 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603a2d8e-2e64-4d45-9449-3e180dd9c340-host\") pod \"crc-debug-g9mq7\" (UID: \"603a2d8e-2e64-4d45-9449-3e180dd9c340\") " pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.445756 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603a2d8e-2e64-4d45-9449-3e180dd9c340-host\") pod \"crc-debug-g9mq7\" (UID: \"603a2d8e-2e64-4d45-9449-3e180dd9c340\") " pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.467895 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gndq8\" (UniqueName: \"kubernetes.io/projected/603a2d8e-2e64-4d45-9449-3e180dd9c340-kube-api-access-gndq8\") pod \"crc-debug-g9mq7\" (UID: \"603a2d8e-2e64-4d45-9449-3e180dd9c340\") " pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.557088 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:20:45 crc kubenswrapper[4744]: I0106 16:20:45.953830 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/crc-debug-g9mq7" event={"ID":"603a2d8e-2e64-4d45-9449-3e180dd9c340","Type":"ContainerStarted","Data":"0c8629f58a56ed8da1e6fd84e23c6b9ba2a7f741574de5d61dfcbe84079c6164"} Jan 06 16:20:58 crc kubenswrapper[4744]: I0106 16:20:58.091432 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/crc-debug-g9mq7" event={"ID":"603a2d8e-2e64-4d45-9449-3e180dd9c340","Type":"ContainerStarted","Data":"949c3e1bbd5fda367b0781011a622105ea2709def5c7db93597c921229a850f2"} Jan 06 16:20:58 crc kubenswrapper[4744]: I0106 16:20:58.192296 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8vxws/crc-debug-g9mq7" podStartSLOduration=1.603758387 podStartE2EDuration="13.192277102s" podCreationTimestamp="2026-01-06 16:20:45 +0000 UTC" firstStartedPulling="2026-01-06 16:20:45.594811588 +0000 UTC m=+6242.222277906" lastFinishedPulling="2026-01-06 16:20:57.183330303 +0000 UTC m=+6253.810796621" observedRunningTime="2026-01-06 16:20:58.183462507 +0000 UTC m=+6254.810928825" watchObservedRunningTime="2026-01-06 16:20:58.192277102 +0000 UTC m=+6254.819743420" Jan 06 16:21:37 crc kubenswrapper[4744]: I0106 16:21:37.501886 4744 generic.go:334] "Generic (PLEG): container finished" podID="603a2d8e-2e64-4d45-9449-3e180dd9c340" containerID="949c3e1bbd5fda367b0781011a622105ea2709def5c7db93597c921229a850f2" exitCode=0 Jan 06 16:21:37 crc kubenswrapper[4744]: I0106 16:21:37.501990 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/crc-debug-g9mq7" event={"ID":"603a2d8e-2e64-4d45-9449-3e180dd9c340","Type":"ContainerDied","Data":"949c3e1bbd5fda367b0781011a622105ea2709def5c7db93597c921229a850f2"} Jan 06 16:21:38 crc kubenswrapper[4744]: I0106 16:21:38.646497 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:21:38 crc kubenswrapper[4744]: I0106 16:21:38.706027 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8vxws/crc-debug-g9mq7"] Jan 06 16:21:38 crc kubenswrapper[4744]: I0106 16:21:38.717237 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8vxws/crc-debug-g9mq7"] Jan 06 16:21:38 crc kubenswrapper[4744]: I0106 16:21:38.756298 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gndq8\" (UniqueName: \"kubernetes.io/projected/603a2d8e-2e64-4d45-9449-3e180dd9c340-kube-api-access-gndq8\") pod \"603a2d8e-2e64-4d45-9449-3e180dd9c340\" (UID: \"603a2d8e-2e64-4d45-9449-3e180dd9c340\") " Jan 06 16:21:38 crc kubenswrapper[4744]: I0106 16:21:38.756551 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603a2d8e-2e64-4d45-9449-3e180dd9c340-host\") pod \"603a2d8e-2e64-4d45-9449-3e180dd9c340\" (UID: \"603a2d8e-2e64-4d45-9449-3e180dd9c340\") " Jan 06 16:21:38 crc kubenswrapper[4744]: I0106 16:21:38.756737 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/603a2d8e-2e64-4d45-9449-3e180dd9c340-host" (OuterVolumeSpecName: "host") pod "603a2d8e-2e64-4d45-9449-3e180dd9c340" (UID: "603a2d8e-2e64-4d45-9449-3e180dd9c340"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 16:21:38 crc kubenswrapper[4744]: I0106 16:21:38.758869 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/603a2d8e-2e64-4d45-9449-3e180dd9c340-host\") on node \"crc\" DevicePath \"\"" Jan 06 16:21:38 crc kubenswrapper[4744]: I0106 16:21:38.762801 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/603a2d8e-2e64-4d45-9449-3e180dd9c340-kube-api-access-gndq8" (OuterVolumeSpecName: "kube-api-access-gndq8") pod "603a2d8e-2e64-4d45-9449-3e180dd9c340" (UID: "603a2d8e-2e64-4d45-9449-3e180dd9c340"). InnerVolumeSpecName "kube-api-access-gndq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:21:38 crc kubenswrapper[4744]: I0106 16:21:38.862300 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gndq8\" (UniqueName: \"kubernetes.io/projected/603a2d8e-2e64-4d45-9449-3e180dd9c340-kube-api-access-gndq8\") on node \"crc\" DevicePath \"\"" Jan 06 16:21:39 crc kubenswrapper[4744]: I0106 16:21:39.525088 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c8629f58a56ed8da1e6fd84e23c6b9ba2a7f741574de5d61dfcbe84079c6164" Jan 06 16:21:39 crc kubenswrapper[4744]: I0106 16:21:39.525111 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-g9mq7" Jan 06 16:21:39 crc kubenswrapper[4744]: I0106 16:21:39.727714 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="603a2d8e-2e64-4d45-9449-3e180dd9c340" path="/var/lib/kubelet/pods/603a2d8e-2e64-4d45-9449-3e180dd9c340/volumes" Jan 06 16:21:39 crc kubenswrapper[4744]: I0106 16:21:39.925530 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8vxws/crc-debug-pg77r"] Jan 06 16:21:39 crc kubenswrapper[4744]: E0106 16:21:39.926101 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603a2d8e-2e64-4d45-9449-3e180dd9c340" containerName="container-00" Jan 06 16:21:39 crc kubenswrapper[4744]: I0106 16:21:39.926123 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="603a2d8e-2e64-4d45-9449-3e180dd9c340" containerName="container-00" Jan 06 16:21:39 crc kubenswrapper[4744]: I0106 16:21:39.926453 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="603a2d8e-2e64-4d45-9449-3e180dd9c340" containerName="container-00" Jan 06 16:21:39 crc kubenswrapper[4744]: I0106 16:21:39.927581 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:39 crc kubenswrapper[4744]: I0106 16:21:39.993605 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6eb049-b64b-426d-889e-9b358ba95146-host\") pod \"crc-debug-pg77r\" (UID: \"4c6eb049-b64b-426d-889e-9b358ba95146\") " pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:39 crc kubenswrapper[4744]: I0106 16:21:39.993739 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gwhq\" (UniqueName: \"kubernetes.io/projected/4c6eb049-b64b-426d-889e-9b358ba95146-kube-api-access-7gwhq\") pod \"crc-debug-pg77r\" (UID: \"4c6eb049-b64b-426d-889e-9b358ba95146\") " pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:40 crc kubenswrapper[4744]: I0106 16:21:40.096207 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6eb049-b64b-426d-889e-9b358ba95146-host\") pod \"crc-debug-pg77r\" (UID: \"4c6eb049-b64b-426d-889e-9b358ba95146\") " pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:40 crc kubenswrapper[4744]: I0106 16:21:40.096337 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gwhq\" (UniqueName: \"kubernetes.io/projected/4c6eb049-b64b-426d-889e-9b358ba95146-kube-api-access-7gwhq\") pod \"crc-debug-pg77r\" (UID: \"4c6eb049-b64b-426d-889e-9b358ba95146\") " pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:40 crc kubenswrapper[4744]: I0106 16:21:40.096385 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6eb049-b64b-426d-889e-9b358ba95146-host\") pod \"crc-debug-pg77r\" (UID: \"4c6eb049-b64b-426d-889e-9b358ba95146\") " pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:40 crc kubenswrapper[4744]: I0106 16:21:40.116611 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gwhq\" (UniqueName: \"kubernetes.io/projected/4c6eb049-b64b-426d-889e-9b358ba95146-kube-api-access-7gwhq\") pod \"crc-debug-pg77r\" (UID: \"4c6eb049-b64b-426d-889e-9b358ba95146\") " pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:40 crc kubenswrapper[4744]: I0106 16:21:40.247661 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:40 crc kubenswrapper[4744]: I0106 16:21:40.539024 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/crc-debug-pg77r" event={"ID":"4c6eb049-b64b-426d-889e-9b358ba95146","Type":"ContainerStarted","Data":"731647393aaf21e28a198dee07cdc90e281b8f6630ce7e4b4c9d8ae859f07bbf"} Jan 06 16:21:41 crc kubenswrapper[4744]: I0106 16:21:41.548859 4744 generic.go:334] "Generic (PLEG): container finished" podID="4c6eb049-b64b-426d-889e-9b358ba95146" containerID="988142daf64e81f40afde54d049e0ba87a0b8d44824ebec2048b209ae22d4911" exitCode=0 Jan 06 16:21:41 crc kubenswrapper[4744]: I0106 16:21:41.548915 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/crc-debug-pg77r" event={"ID":"4c6eb049-b64b-426d-889e-9b358ba95146","Type":"ContainerDied","Data":"988142daf64e81f40afde54d049e0ba87a0b8d44824ebec2048b209ae22d4911"} Jan 06 16:21:41 crc kubenswrapper[4744]: I0106 16:21:41.900915 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8vxws/crc-debug-pg77r"] Jan 06 16:21:41 crc kubenswrapper[4744]: I0106 16:21:41.930078 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8vxws/crc-debug-pg77r"] Jan 06 16:21:42 crc kubenswrapper[4744]: I0106 16:21:42.682206 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:42 crc kubenswrapper[4744]: I0106 16:21:42.766971 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gwhq\" (UniqueName: \"kubernetes.io/projected/4c6eb049-b64b-426d-889e-9b358ba95146-kube-api-access-7gwhq\") pod \"4c6eb049-b64b-426d-889e-9b358ba95146\" (UID: \"4c6eb049-b64b-426d-889e-9b358ba95146\") " Jan 06 16:21:42 crc kubenswrapper[4744]: I0106 16:21:42.767045 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6eb049-b64b-426d-889e-9b358ba95146-host\") pod \"4c6eb049-b64b-426d-889e-9b358ba95146\" (UID: \"4c6eb049-b64b-426d-889e-9b358ba95146\") " Jan 06 16:21:42 crc kubenswrapper[4744]: I0106 16:21:42.767239 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c6eb049-b64b-426d-889e-9b358ba95146-host" (OuterVolumeSpecName: "host") pod "4c6eb049-b64b-426d-889e-9b358ba95146" (UID: "4c6eb049-b64b-426d-889e-9b358ba95146"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 16:21:42 crc kubenswrapper[4744]: I0106 16:21:42.767850 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4c6eb049-b64b-426d-889e-9b358ba95146-host\") on node \"crc\" DevicePath \"\"" Jan 06 16:21:42 crc kubenswrapper[4744]: I0106 16:21:42.774244 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c6eb049-b64b-426d-889e-9b358ba95146-kube-api-access-7gwhq" (OuterVolumeSpecName: "kube-api-access-7gwhq") pod "4c6eb049-b64b-426d-889e-9b358ba95146" (UID: "4c6eb049-b64b-426d-889e-9b358ba95146"). InnerVolumeSpecName "kube-api-access-7gwhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:21:42 crc kubenswrapper[4744]: I0106 16:21:42.869980 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gwhq\" (UniqueName: \"kubernetes.io/projected/4c6eb049-b64b-426d-889e-9b358ba95146-kube-api-access-7gwhq\") on node \"crc\" DevicePath \"\"" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.050337 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8vxws/crc-debug-cpv5v"] Jan 06 16:21:43 crc kubenswrapper[4744]: E0106 16:21:43.050779 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c6eb049-b64b-426d-889e-9b358ba95146" containerName="container-00" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.050796 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c6eb049-b64b-426d-889e-9b358ba95146" containerName="container-00" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.051040 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c6eb049-b64b-426d-889e-9b358ba95146" containerName="container-00" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.051849 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.176268 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1f10ad1-77cb-4750-937b-0b6004fbac94-host\") pod \"crc-debug-cpv5v\" (UID: \"d1f10ad1-77cb-4750-937b-0b6004fbac94\") " pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.176523 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcgkl\" (UniqueName: \"kubernetes.io/projected/d1f10ad1-77cb-4750-937b-0b6004fbac94-kube-api-access-rcgkl\") pod \"crc-debug-cpv5v\" (UID: \"d1f10ad1-77cb-4750-937b-0b6004fbac94\") " pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.278578 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcgkl\" (UniqueName: \"kubernetes.io/projected/d1f10ad1-77cb-4750-937b-0b6004fbac94-kube-api-access-rcgkl\") pod \"crc-debug-cpv5v\" (UID: \"d1f10ad1-77cb-4750-937b-0b6004fbac94\") " pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.278721 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1f10ad1-77cb-4750-937b-0b6004fbac94-host\") pod \"crc-debug-cpv5v\" (UID: \"d1f10ad1-77cb-4750-937b-0b6004fbac94\") " pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.278828 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1f10ad1-77cb-4750-937b-0b6004fbac94-host\") pod \"crc-debug-cpv5v\" (UID: \"d1f10ad1-77cb-4750-937b-0b6004fbac94\") " pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.297935 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcgkl\" (UniqueName: \"kubernetes.io/projected/d1f10ad1-77cb-4750-937b-0b6004fbac94-kube-api-access-rcgkl\") pod \"crc-debug-cpv5v\" (UID: \"d1f10ad1-77cb-4750-937b-0b6004fbac94\") " pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.378391 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.577503 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="731647393aaf21e28a198dee07cdc90e281b8f6630ce7e4b4c9d8ae859f07bbf" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.578019 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-pg77r" Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.579837 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/crc-debug-cpv5v" event={"ID":"d1f10ad1-77cb-4750-937b-0b6004fbac94","Type":"ContainerStarted","Data":"945843f4a6911ae71152b38a8f8533fb406e471caf0ee003726cb39babb43b56"} Jan 06 16:21:43 crc kubenswrapper[4744]: I0106 16:21:43.728699 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c6eb049-b64b-426d-889e-9b358ba95146" path="/var/lib/kubelet/pods/4c6eb049-b64b-426d-889e-9b358ba95146/volumes" Jan 06 16:21:44 crc kubenswrapper[4744]: I0106 16:21:44.593724 4744 generic.go:334] "Generic (PLEG): container finished" podID="d1f10ad1-77cb-4750-937b-0b6004fbac94" containerID="6391d5afab4d1635389d4f2cddad5bf92d9951c76c70e3e80ffc84bfa88b9e68" exitCode=0 Jan 06 16:21:44 crc kubenswrapper[4744]: I0106 16:21:44.593864 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/crc-debug-cpv5v" event={"ID":"d1f10ad1-77cb-4750-937b-0b6004fbac94","Type":"ContainerDied","Data":"6391d5afab4d1635389d4f2cddad5bf92d9951c76c70e3e80ffc84bfa88b9e68"} Jan 06 16:21:44 crc kubenswrapper[4744]: I0106 16:21:44.635259 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8vxws/crc-debug-cpv5v"] Jan 06 16:21:44 crc kubenswrapper[4744]: I0106 16:21:44.648993 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8vxws/crc-debug-cpv5v"] Jan 06 16:21:45 crc kubenswrapper[4744]: I0106 16:21:45.732003 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:45 crc kubenswrapper[4744]: I0106 16:21:45.835832 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1f10ad1-77cb-4750-937b-0b6004fbac94-host\") pod \"d1f10ad1-77cb-4750-937b-0b6004fbac94\" (UID: \"d1f10ad1-77cb-4750-937b-0b6004fbac94\") " Jan 06 16:21:45 crc kubenswrapper[4744]: I0106 16:21:45.835904 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcgkl\" (UniqueName: \"kubernetes.io/projected/d1f10ad1-77cb-4750-937b-0b6004fbac94-kube-api-access-rcgkl\") pod \"d1f10ad1-77cb-4750-937b-0b6004fbac94\" (UID: \"d1f10ad1-77cb-4750-937b-0b6004fbac94\") " Jan 06 16:21:45 crc kubenswrapper[4744]: I0106 16:21:45.835968 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1f10ad1-77cb-4750-937b-0b6004fbac94-host" (OuterVolumeSpecName: "host") pod "d1f10ad1-77cb-4750-937b-0b6004fbac94" (UID: "d1f10ad1-77cb-4750-937b-0b6004fbac94"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 06 16:21:45 crc kubenswrapper[4744]: I0106 16:21:45.836563 4744 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d1f10ad1-77cb-4750-937b-0b6004fbac94-host\") on node \"crc\" DevicePath \"\"" Jan 06 16:21:45 crc kubenswrapper[4744]: I0106 16:21:45.843266 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f10ad1-77cb-4750-937b-0b6004fbac94-kube-api-access-rcgkl" (OuterVolumeSpecName: "kube-api-access-rcgkl") pod "d1f10ad1-77cb-4750-937b-0b6004fbac94" (UID: "d1f10ad1-77cb-4750-937b-0b6004fbac94"). InnerVolumeSpecName "kube-api-access-rcgkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:21:45 crc kubenswrapper[4744]: I0106 16:21:45.938468 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcgkl\" (UniqueName: \"kubernetes.io/projected/d1f10ad1-77cb-4750-937b-0b6004fbac94-kube-api-access-rcgkl\") on node \"crc\" DevicePath \"\"" Jan 06 16:21:46 crc kubenswrapper[4744]: I0106 16:21:46.628930 4744 scope.go:117] "RemoveContainer" containerID="6391d5afab4d1635389d4f2cddad5bf92d9951c76c70e3e80ffc84bfa88b9e68" Jan 06 16:21:46 crc kubenswrapper[4744]: I0106 16:21:46.629009 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/crc-debug-cpv5v" Jan 06 16:21:47 crc kubenswrapper[4744]: I0106 16:21:47.726525 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1f10ad1-77cb-4750-937b-0b6004fbac94" path="/var/lib/kubelet/pods/d1f10ad1-77cb-4750-937b-0b6004fbac94/volumes" Jan 06 16:22:16 crc kubenswrapper[4744]: I0106 16:22:16.905283 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f/aodh-api/0.log" Jan 06 16:22:17 crc kubenswrapper[4744]: I0106 16:22:17.557401 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f/aodh-listener/0.log" Jan 06 16:22:17 crc kubenswrapper[4744]: I0106 16:22:17.557447 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f/aodh-notifier/0.log" Jan 06 16:22:17 crc kubenswrapper[4744]: I0106 16:22:17.580740 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_d77ef1d6-bd13-43f6-ac08-3da1f1a20d4f/aodh-evaluator/0.log" Jan 06 16:22:17 crc kubenswrapper[4744]: I0106 16:22:17.780173 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b8c67fcd6-nlf9f_c2c0e750-af72-4679-acae-81c93ac6b40e/barbican-api-log/0.log" Jan 06 16:22:17 crc kubenswrapper[4744]: I0106 16:22:17.790698 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6b8c67fcd6-nlf9f_c2c0e750-af72-4679-acae-81c93ac6b40e/barbican-api/0.log" Jan 06 16:22:17 crc kubenswrapper[4744]: I0106 16:22:17.881680 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5d48fc6d-cr8dv_762cdea3-aba0-4f8f-96c8-4dfaab8bf92a/barbican-keystone-listener/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.099689 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-74dd54d475-wqvpn_bc804009-f9fd-4d70-bdd8-c6c61a544549/barbican-worker/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.099756 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5d48fc6d-cr8dv_762cdea3-aba0-4f8f-96c8-4dfaab8bf92a/barbican-keystone-listener-log/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.122897 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-74dd54d475-wqvpn_bc804009-f9fd-4d70-bdd8-c6c61a544549/barbican-worker-log/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.356749 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-wqrwz_eac03a71-e859-4391-b58e-f4dc4fdfa0ce/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.422745 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_32ce34b0-736e-4d20-9c3b-4c76656092dc/ceilometer-central-agent/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.597062 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_32ce34b0-736e-4d20-9c3b-4c76656092dc/sg-core/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.619898 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_32ce34b0-736e-4d20-9c3b-4c76656092dc/proxy-httpd/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.635417 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_32ce34b0-736e-4d20-9c3b-4c76656092dc/ceilometer-notification-agent/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.851721 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b/cinder-api-log/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.917643 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b9fd595c-54f3-43e4-bea9-7b5b4a9b4b0b/cinder-api/0.log" Jan 06 16:22:18 crc kubenswrapper[4744]: I0106 16:22:18.923299 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_1ad5afd6-6681-4bae-a1e7-a9a7a27dae40/cinder-scheduler/0.log" Jan 06 16:22:19 crc kubenswrapper[4744]: I0106 16:22:19.089600 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_1ad5afd6-6681-4bae-a1e7-a9a7a27dae40/probe/0.log" Jan 06 16:22:19 crc kubenswrapper[4744]: I0106 16:22:19.225701 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-cqbx9_ae6684c3-9c28-4aca-a8c3-7a4545d6eb55/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:19 crc kubenswrapper[4744]: I0106 16:22:19.316308 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-2n6h4_2b9ab759-a383-4ec8-ad5a-578f35de2a5e/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:19 crc kubenswrapper[4744]: I0106 16:22:19.473898 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-nb6pm_aa8be235-0419-4ff3-8e76-d48ddbcfe47f/init/0.log" Jan 06 16:22:19 crc kubenswrapper[4744]: I0106 16:22:19.633675 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-nb6pm_aa8be235-0419-4ff3-8e76-d48ddbcfe47f/init/0.log" Jan 06 16:22:19 crc kubenswrapper[4744]: I0106 16:22:19.677584 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5596c69fcc-nb6pm_aa8be235-0419-4ff3-8e76-d48ddbcfe47f/dnsmasq-dns/0.log" Jan 06 16:22:19 crc kubenswrapper[4744]: I0106 16:22:19.749215 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-sgbd5_98606045-db2f-49f7-a0a3-c88b391364c6/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:19 crc kubenswrapper[4744]: I0106 16:22:19.977254 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ef75f535-8a96-40b4-9719-965a1c97a9be/glance-log/0.log" Jan 06 16:22:19 crc kubenswrapper[4744]: I0106 16:22:19.978409 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ef75f535-8a96-40b4-9719-965a1c97a9be/glance-httpd/0.log" Jan 06 16:22:20 crc kubenswrapper[4744]: I0106 16:22:20.123540 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6ae7b1ee-3df3-443f-92e9-c4a68e6829e3/glance-httpd/0.log" Jan 06 16:22:20 crc kubenswrapper[4744]: I0106 16:22:20.226369 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6ae7b1ee-3df3-443f-92e9-c4a68e6829e3/glance-log/0.log" Jan 06 16:22:20 crc kubenswrapper[4744]: I0106 16:22:20.847331 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-5c5c49fd76-ww78j_39cc6313-b52f-4b01-b593-41008188e011/heat-engine/0.log" Jan 06 16:22:20 crc kubenswrapper[4744]: I0106 16:22:20.904624 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-db66db7b6-ghcg8_0a70b52f-ba28-4e3d-ba4c-0e41cd023bf0/heat-api/0.log" Jan 06 16:22:21 crc kubenswrapper[4744]: I0106 16:22:21.099782 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-8fcb46fc5-bk6ql_fd669f35-602e-4680-a4ec-2d9bb7e66fca/heat-cfnapi/0.log" Jan 06 16:22:21 crc kubenswrapper[4744]: I0106 16:22:21.161532 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-rh4vz_5f9705e7-7499-4378-8d58-c9373c83e5ed/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:21 crc kubenswrapper[4744]: I0106 16:22:21.170750 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-jb2sb_f50fabe8-fbbb-406f-9786-ff32ff95feaa/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:21 crc kubenswrapper[4744]: I0106 16:22:21.600558 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29461861-vjkz8_d3548869-d949-4d99-8968-8cdf8b347771/keystone-cron/0.log" Jan 06 16:22:21 crc kubenswrapper[4744]: I0106 16:22:21.683243 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-77f58bbbd5-m6wzx_6cee76ef-12c9-4949-aeeb-faea5a0f5ddb/keystone-api/0.log" Jan 06 16:22:21 crc kubenswrapper[4744]: I0106 16:22:21.709021 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29461921-5vhv9_2da36365-1bbd-45e9-bd2f-392f7b34b25d/keystone-cron/0.log" Jan 06 16:22:21 crc kubenswrapper[4744]: I0106 16:22:21.896822 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-lpxf4_5d8bb7f8-6093-4f4a-81a1-6917f83346bf/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:21 crc kubenswrapper[4744]: I0106 16:22:21.899091 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_a9ce0f2f-e4c5-4cc5-8196-0340d738f830/kube-state-metrics/0.log" Jan 06 16:22:22 crc kubenswrapper[4744]: I0106 16:22:22.053023 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-pp45w_362fd3ee-198d-4233-b757-8db8895dc05b/logging-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:22 crc kubenswrapper[4744]: I0106 16:22:22.308989 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_6d3de113-d71d-4f12-bebc-d3466483e800/mysqld-exporter/0.log" Jan 06 16:22:22 crc kubenswrapper[4744]: I0106 16:22:22.514044 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c5f6db8dc-vwrjb_ab974dd7-0476-4bec-8c0d-b7bdcf52470f/neutron-httpd/0.log" Jan 06 16:22:22 crc kubenswrapper[4744]: I0106 16:22:22.514793 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c5f6db8dc-vwrjb_ab974dd7-0476-4bec-8c0d-b7bdcf52470f/neutron-api/0.log" Jan 06 16:22:22 crc kubenswrapper[4744]: I0106 16:22:22.608847 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-5m49j_c913d43d-9615-4436-ae90-56c68245d4a6/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:23 crc kubenswrapper[4744]: I0106 16:22:23.082651 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_46daf908-4aec-483b-ba31-ee145e42de95/nova-api-log/0.log" Jan 06 16:22:23 crc kubenswrapper[4744]: I0106 16:22:23.267474 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_2a76b16a-3d97-45e9-aefe-3de95463d5fd/nova-cell0-conductor-conductor/0.log" Jan 06 16:22:23 crc kubenswrapper[4744]: I0106 16:22:23.413099 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0e274934-d185-4e44-b1a1-46d78395174c/nova-cell1-conductor-conductor/0.log" Jan 06 16:22:23 crc kubenswrapper[4744]: I0106 16:22:23.581007 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_1999f288-0cad-400d-9c96-f073c424881b/nova-cell1-novncproxy-novncproxy/0.log" Jan 06 16:22:23 crc kubenswrapper[4744]: I0106 16:22:23.585423 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_46daf908-4aec-483b-ba31-ee145e42de95/nova-api-api/0.log" Jan 06 16:22:23 crc kubenswrapper[4744]: I0106 16:22:23.699558 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-kdmcr_af287382-0a8b-4a8c-a3a3-fbb41c6da499/nova-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:23 crc kubenswrapper[4744]: I0106 16:22:23.925307 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1af5c39c-9101-4d0e-85be-a4556aef4cbb/nova-metadata-log/0.log" Jan 06 16:22:24 crc kubenswrapper[4744]: I0106 16:22:24.207339 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bc13577b-cb8c-4e94-9109-34d00a75d148/mysql-bootstrap/0.log" Jan 06 16:22:24 crc kubenswrapper[4744]: I0106 16:22:24.315486 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8221b070-0d54-439f-887a-9a401d634bea/nova-scheduler-scheduler/0.log" Jan 06 16:22:24 crc kubenswrapper[4744]: I0106 16:22:24.414150 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bc13577b-cb8c-4e94-9109-34d00a75d148/mysql-bootstrap/0.log" Jan 06 16:22:24 crc kubenswrapper[4744]: I0106 16:22:24.460892 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_bc13577b-cb8c-4e94-9109-34d00a75d148/galera/0.log" Jan 06 16:22:24 crc kubenswrapper[4744]: I0106 16:22:24.892238 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b2b4c15b-189d-41b2-b121-fdc25beb18a4/mysql-bootstrap/0.log" Jan 06 16:22:25 crc kubenswrapper[4744]: I0106 16:22:25.057505 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b2b4c15b-189d-41b2-b121-fdc25beb18a4/mysql-bootstrap/0.log" Jan 06 16:22:25 crc kubenswrapper[4744]: I0106 16:22:25.119501 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_b2b4c15b-189d-41b2-b121-fdc25beb18a4/galera/0.log" Jan 06 16:22:25 crc kubenswrapper[4744]: I0106 16:22:25.299499 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_50332b7c-a1e8-4cd4-95d6-463173669507/openstackclient/0.log" Jan 06 16:22:25 crc kubenswrapper[4744]: I0106 16:22:25.389024 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-nsnv2_7e6b0b81-7dd2-4030-827e-54bbc1a59635/openstack-network-exporter/0.log" Jan 06 16:22:25 crc kubenswrapper[4744]: I0106 16:22:25.578412 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qcph6_78600e4f-de4a-4079-b849-b7b2319685b0/ovsdb-server-init/0.log" Jan 06 16:22:25 crc kubenswrapper[4744]: I0106 16:22:25.769836 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qcph6_78600e4f-de4a-4079-b849-b7b2319685b0/ovsdb-server-init/0.log" Jan 06 16:22:25 crc kubenswrapper[4744]: I0106 16:22:25.840706 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qcph6_78600e4f-de4a-4079-b849-b7b2319685b0/ovsdb-server/0.log" Jan 06 16:22:25 crc kubenswrapper[4744]: I0106 16:22:25.883276 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-qcph6_78600e4f-de4a-4079-b849-b7b2319685b0/ovs-vswitchd/0.log" Jan 06 16:22:26 crc kubenswrapper[4744]: I0106 16:22:26.062659 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-r4r76_7f86f187-37db-488e-b6f6-22a5becacecd/ovn-controller/0.log" Jan 06 16:22:26 crc kubenswrapper[4744]: I0106 16:22:26.294505 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1af5c39c-9101-4d0e-85be-a4556aef4cbb/nova-metadata-metadata/0.log" Jan 06 16:22:26 crc kubenswrapper[4744]: I0106 16:22:26.298936 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-twr7f_fcdb07a4-62d3-4425-9319-4e58bcb6831c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:26 crc kubenswrapper[4744]: I0106 16:22:26.457679 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_73416f0d-1377-4ac4-ab12-c5d54151325b/openstack-network-exporter/0.log" Jan 06 16:22:26 crc kubenswrapper[4744]: I0106 16:22:26.563338 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_73416f0d-1377-4ac4-ab12-c5d54151325b/ovn-northd/0.log" Jan 06 16:22:26 crc kubenswrapper[4744]: I0106 16:22:26.612868 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ab7e8638-3ee7-472b-a6de-a15b321e5fd5/openstack-network-exporter/0.log" Jan 06 16:22:26 crc kubenswrapper[4744]: I0106 16:22:26.673659 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ab7e8638-3ee7-472b-a6de-a15b321e5fd5/ovsdbserver-nb/0.log" Jan 06 16:22:26 crc kubenswrapper[4744]: I0106 16:22:26.832673 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_13c79025-4c6b-4040-9f3d-4b43390b9c17/openstack-network-exporter/0.log" Jan 06 16:22:26 crc kubenswrapper[4744]: I0106 16:22:26.842989 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_13c79025-4c6b-4040-9f3d-4b43390b9c17/ovsdbserver-sb/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.093194 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68984dbdcb-mkv9c_7d48c33e-540e-4bd3-b88a-592dff511604/placement-api/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.135130 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-68984dbdcb-mkv9c_7d48c33e-540e-4bd3-b88a-592dff511604/placement-log/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.162834 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf/init-config-reloader/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.353655 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf/init-config-reloader/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.390339 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf/config-reloader/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.430325 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf/thanos-sidecar/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.439221 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_0ae9e83f-2f67-419c-9929-cfd5bdcfcdbf/prometheus/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.595397 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a33bb3f6-57ba-4e75-af0b-f9895afb79aa/setup-container/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.760463 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a33bb3f6-57ba-4e75-af0b-f9895afb79aa/setup-container/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.817883 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a33bb3f6-57ba-4e75-af0b-f9895afb79aa/rabbitmq/0.log" Jan 06 16:22:27 crc kubenswrapper[4744]: I0106 16:22:27.883669 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b4c6a122-8746-4f2b-a7f9-e3df62969170/setup-container/0.log" Jan 06 16:22:28 crc kubenswrapper[4744]: I0106 16:22:28.335725 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-1_c38b64df-d42d-48ba-a5eb-12f0523a8555/setup-container/0.log" Jan 06 16:22:28 crc kubenswrapper[4744]: I0106 16:22:28.414062 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b4c6a122-8746-4f2b-a7f9-e3df62969170/setup-container/0.log" Jan 06 16:22:28 crc kubenswrapper[4744]: I0106 16:22:28.461531 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b4c6a122-8746-4f2b-a7f9-e3df62969170/rabbitmq/0.log" Jan 06 16:22:28 crc kubenswrapper[4744]: I0106 16:22:28.616828 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-1_c38b64df-d42d-48ba-a5eb-12f0523a8555/setup-container/0.log" Jan 06 16:22:28 crc kubenswrapper[4744]: I0106 16:22:28.672222 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-1_c38b64df-d42d-48ba-a5eb-12f0523a8555/rabbitmq/0.log" Jan 06 16:22:28 crc kubenswrapper[4744]: I0106 16:22:28.756714 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-2_b4c3d9bb-3170-4372-8cdd-ec1bfe092237/setup-container/0.log" Jan 06 16:22:28 crc kubenswrapper[4744]: I0106 16:22:28.927263 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-2_b4c3d9bb-3170-4372-8cdd-ec1bfe092237/setup-container/0.log" Jan 06 16:22:28 crc kubenswrapper[4744]: I0106 16:22:28.954432 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-2_b4c3d9bb-3170-4372-8cdd-ec1bfe092237/rabbitmq/0.log" Jan 06 16:22:29 crc kubenswrapper[4744]: I0106 16:22:29.066644 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-2z9wm_7bb6bb0d-d95e-48e5-96a0-907532d387d2/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:29 crc kubenswrapper[4744]: I0106 16:22:29.247972 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-x5b5w_8664eee2-4c8b-4e3c-9bdd-26658aa1ab76/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:29 crc kubenswrapper[4744]: I0106 16:22:29.279533 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-qm89h_42eceb98-5fea-4ae7-827f-94b867781272/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:29 crc kubenswrapper[4744]: I0106 16:22:29.449249 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-clw5b_95e3d07e-c41f-4d70-8890-6b87c90468b0/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:29 crc kubenswrapper[4744]: I0106 16:22:29.509980 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-km9j8_e77857ea-77e9-4d71-80bb-01107b6b3820/ssh-known-hosts-edpm-deployment/0.log" Jan 06 16:22:29 crc kubenswrapper[4744]: I0106 16:22:29.736332 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-787d958d8c-44gnt_e1974504-5e17-4e84-95bd-7d66ed033ccc/proxy-server/0.log" Jan 06 16:22:29 crc kubenswrapper[4744]: I0106 16:22:29.900778 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-787d958d8c-44gnt_e1974504-5e17-4e84-95bd-7d66ed033ccc/proxy-httpd/0.log" Jan 06 16:22:29 crc kubenswrapper[4744]: I0106 16:22:29.906668 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-fq864_82b9e690-d1d7-473d-99d1-20be1130dfd3/swift-ring-rebalance/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.032671 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/account-auditor/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.122007 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/account-reaper/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.235652 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/account-replicator/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.292949 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/container-auditor/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.296919 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/account-server/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.475337 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/container-server/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.512717 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/object-auditor/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.517109 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/container-updater/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.528662 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/container-replicator/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.681075 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/object-expirer/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.700117 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/object-server/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.766967 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/object-updater/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.779006 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/object-replicator/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.872295 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/rsync/0.log" Jan 06 16:22:30 crc kubenswrapper[4744]: I0106 16:22:30.933739 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c603cde3-b7b0-4851-99df-2abb0aa6b1c4/swift-recon-cron/0.log" Jan 06 16:22:31 crc kubenswrapper[4744]: I0106 16:22:31.102938 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-b2k4d_62caf07e-9021-49bb-9e27-79df992fc4c5/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:31 crc kubenswrapper[4744]: I0106 16:22:31.234746 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-fh4cr_5216d2ec-d2e3-4f91-97d5-5839eaeb173d/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:31 crc kubenswrapper[4744]: I0106 16:22:31.537195 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_f28b7936-8bd1-41de-adba-74506e7df59e/memcached/0.log" Jan 06 16:22:31 crc kubenswrapper[4744]: I0106 16:22:31.547054 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-jfbv9_f0380bc6-d687-41af-a5d5-6569a6c299d9/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 06 16:22:44 crc kubenswrapper[4744]: I0106 16:22:44.423982 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:22:44 crc kubenswrapper[4744]: I0106 16:22:44.424441 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:22:56 crc kubenswrapper[4744]: I0106 16:22:56.792455 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-f6f74d6db-lhppk_130c27ae-5c64-4c06-b5e8-9a1aacea630e/manager/0.log" Jan 06 16:22:56 crc kubenswrapper[4744]: I0106 16:22:56.911989 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq_f1b99747-da20-401e-a710-13ea2d33e7fc/util/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.126514 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq_f1b99747-da20-401e-a710-13ea2d33e7fc/util/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.160765 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq_f1b99747-da20-401e-a710-13ea2d33e7fc/pull/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.162675 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq_f1b99747-da20-401e-a710-13ea2d33e7fc/pull/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.343419 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq_f1b99747-da20-401e-a710-13ea2d33e7fc/util/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.367991 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq_f1b99747-da20-401e-a710-13ea2d33e7fc/extract/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.396477 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cebc2442dab0e6d5502ae6f2b01a6a2196de54366302cc28bdd41ca215btcsq_f1b99747-da20-401e-a710-13ea2d33e7fc/pull/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.567360 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-78979fc445-bm4m2_5b589b5b-5f62-4e3d-808b-8c404d1432ac/manager/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.582442 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66f8b87655-j8rsd_3653611e-a619-401f-a964-70d245bb4957/manager/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.806315 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7b549fc966-clgkb_8035e88d-2a38-4e95-b204-0dc00fc57bfc/manager/0.log" Jan 06 16:22:57 crc kubenswrapper[4744]: I0106 16:22:57.909873 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-658dd65b86-pnxwz_8421b8a1-9929-43ab-b203-9899c52d2d41/manager/0.log" Jan 06 16:22:58 crc kubenswrapper[4744]: I0106 16:22:58.030546 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7f5ddd8d7b-q9k7b_f1558cc4-7188-42b8-858a-e83606895217/manager/0.log" Jan 06 16:22:58 crc kubenswrapper[4744]: I0106 16:22:58.218148 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-f99f54bc8-hwxtd_aabafe79-ea6f-4f01-9efb-e1c1a9e2334e/manager/0.log" Jan 06 16:22:58 crc kubenswrapper[4744]: I0106 16:22:58.473301 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6d99759cf-pgssl_749a8719-1599-4225-a5f1-165fbe644aca/manager/0.log" Jan 06 16:22:58 crc kubenswrapper[4744]: I0106 16:22:58.503326 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-568985c78-dhg75_66470ebb-8a02-4745-941b-5b6a1b97df1c/manager/0.log" Jan 06 16:22:58 crc kubenswrapper[4744]: I0106 16:22:58.583209 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-598945d5b8-qc2xd_8cc57759-7dc1-4360-8d81-4744355ec4df/manager/0.log" Jan 06 16:22:58 crc kubenswrapper[4744]: I0106 16:22:58.707561 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-7b88bfc995-k8n6t_0abaa053-82ac-4b9f-a4d2-2bc5024a97eb/manager/0.log" Jan 06 16:22:58 crc kubenswrapper[4744]: I0106 16:22:58.872231 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7cd87b778f-bl6hr_5a327dca-6d91-4a3d-a824-b9f9080338dc/manager/0.log" Jan 06 16:22:58 crc kubenswrapper[4744]: I0106 16:22:58.997583 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5fbbf8b6cc-twhxx_70440288-e0e0-42fd-b310-da1267956b3d/manager/0.log" Jan 06 16:22:59 crc kubenswrapper[4744]: I0106 16:22:59.085083 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-68c649d9d-dzwnd_96fdaf1c-2733-4272-abda-fb28fa6f6a6a/manager/0.log" Jan 06 16:22:59 crc kubenswrapper[4744]: I0106 16:22:59.131224 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-78948ddfd785mfz_88292fa9-b45e-4681-97df-e05a423895cc/manager/0.log" Jan 06 16:22:59 crc kubenswrapper[4744]: I0106 16:22:59.535387 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-7hrfh_458bdba1-55cb-49ad-81ef-5e4c8b07c920/registry-server/0.log" Jan 06 16:22:59 crc kubenswrapper[4744]: I0106 16:22:59.543212 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-749c9566f9-kw8ht_7be82a54-010f-4102-bea3-f3ce988e65b6/operator/0.log" Jan 06 16:22:59 crc kubenswrapper[4744]: I0106 16:22:59.825362 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bf6d4f946-8nnk9_36725987-d3b2-4216-bfe0-69fef2d8bfd7/manager/0.log" Jan 06 16:22:59 crc kubenswrapper[4744]: I0106 16:22:59.961374 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-9b6f8f78c-fvltb_9ef85238-0ee3-44af-98ad-734994758fec/manager/0.log" Jan 06 16:23:00 crc kubenswrapper[4744]: I0106 16:23:00.075533 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-bjtfr_fec21671-aa20-4019-bc89-9e0fc135e394/operator/0.log" Jan 06 16:23:00 crc kubenswrapper[4744]: I0106 16:23:00.291255 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-bb586bbf4-p5kz7_73e35a67-364a-487c-a7bf-5e537cbbecbe/manager/0.log" Jan 06 16:23:00 crc kubenswrapper[4744]: I0106 16:23:00.544941 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6c866cfdcb-8q8ks_65ec28e9-38c5-4972-b8b8-1447de87ee85/manager/0.log" Jan 06 16:23:00 crc kubenswrapper[4744]: I0106 16:23:00.742897 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-9dbdf6486-9j2kx_1ea5e62f-608a-4fa7-8887-691c1c314a6e/manager/0.log" Jan 06 16:23:00 crc kubenswrapper[4744]: I0106 16:23:00.788443 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5459d89946-2qz6t_7f9fe9dd-bede-4261-9a8a-d23812cffca2/manager/0.log" Jan 06 16:23:00 crc kubenswrapper[4744]: I0106 16:23:00.820830 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5649998c8c-g454h_d18584e4-6488-4fc3-9992-9ded578fd05e/manager/0.log" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.371154 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fxk5l"] Jan 06 16:23:11 crc kubenswrapper[4744]: E0106 16:23:11.372110 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f10ad1-77cb-4750-937b-0b6004fbac94" containerName="container-00" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.372123 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f10ad1-77cb-4750-937b-0b6004fbac94" containerName="container-00" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.372352 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f10ad1-77cb-4750-937b-0b6004fbac94" containerName="container-00" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.374873 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.387533 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fxk5l"] Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.476397 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-catalog-content\") pod \"certified-operators-fxk5l\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.476583 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h6sb\" (UniqueName: \"kubernetes.io/projected/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-kube-api-access-7h6sb\") pod \"certified-operators-fxk5l\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.476695 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-utilities\") pod \"certified-operators-fxk5l\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.578811 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h6sb\" (UniqueName: \"kubernetes.io/projected/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-kube-api-access-7h6sb\") pod \"certified-operators-fxk5l\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.578910 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-utilities\") pod \"certified-operators-fxk5l\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.579010 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-catalog-content\") pod \"certified-operators-fxk5l\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.579453 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-catalog-content\") pod \"certified-operators-fxk5l\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.580031 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-utilities\") pod \"certified-operators-fxk5l\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.597365 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h6sb\" (UniqueName: \"kubernetes.io/projected/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-kube-api-access-7h6sb\") pod \"certified-operators-fxk5l\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:11 crc kubenswrapper[4744]: I0106 16:23:11.699403 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:12 crc kubenswrapper[4744]: I0106 16:23:12.210223 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fxk5l"] Jan 06 16:23:12 crc kubenswrapper[4744]: I0106 16:23:12.592842 4744 generic.go:334] "Generic (PLEG): container finished" podID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerID="c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c" exitCode=0 Jan 06 16:23:12 crc kubenswrapper[4744]: I0106 16:23:12.593138 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxk5l" event={"ID":"25f3e669-4a9e-48cd-be1c-50e76ecdef1f","Type":"ContainerDied","Data":"c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c"} Jan 06 16:23:12 crc kubenswrapper[4744]: I0106 16:23:12.594389 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxk5l" event={"ID":"25f3e669-4a9e-48cd-be1c-50e76ecdef1f","Type":"ContainerStarted","Data":"ded4179d7afcd0d24c9f3d90abcb7b7180a2f8ac78d42f98eceaf962f0dca51e"} Jan 06 16:23:12 crc kubenswrapper[4744]: I0106 16:23:12.595587 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 16:23:14 crc kubenswrapper[4744]: I0106 16:23:14.424391 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:23:14 crc kubenswrapper[4744]: I0106 16:23:14.424882 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:23:20 crc kubenswrapper[4744]: I0106 16:23:20.754608 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-gdt2p_ed5d8f62-5fc4-46f9-bc8b-0154549a9222/control-plane-machine-set-operator/0.log" Jan 06 16:23:20 crc kubenswrapper[4744]: I0106 16:23:20.899778 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mpb2d_5e458d23-aef6-4e91-9741-46d74c403131/kube-rbac-proxy/0.log" Jan 06 16:23:20 crc kubenswrapper[4744]: I0106 16:23:20.947675 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mpb2d_5e458d23-aef6-4e91-9741-46d74c403131/machine-api-operator/0.log" Jan 06 16:23:34 crc kubenswrapper[4744]: I0106 16:23:34.994647 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-9gjwj_c472e993-f700-40e9-9a30-6d81b169cd5c/cert-manager-controller/0.log" Jan 06 16:23:35 crc kubenswrapper[4744]: I0106 16:23:35.178638 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-tvhp5_9539eba1-94db-41e9-80ed-ba1de1978f94/cert-manager-webhook/0.log" Jan 06 16:23:35 crc kubenswrapper[4744]: I0106 16:23:35.432778 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-66g5r_4947ffd9-d5ba-476d-89b8-c60b573f436d/cert-manager-cainjector/0.log" Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.423336 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.424108 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.424149 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.424993 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.425049 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" gracePeriod=600 Jan 06 16:23:44 crc kubenswrapper[4744]: E0106 16:23:44.547420 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.960111 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxk5l" event={"ID":"25f3e669-4a9e-48cd-be1c-50e76ecdef1f","Type":"ContainerStarted","Data":"7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44"} Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.963242 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" exitCode=0 Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.963288 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f"} Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.963322 4744 scope.go:117] "RemoveContainer" containerID="221b2e5281ccbb5da1e0a7b76c3b63ae8d2ef757f93ecb10652f51c87dfd189b" Jan 06 16:23:44 crc kubenswrapper[4744]: I0106 16:23:44.963931 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:23:44 crc kubenswrapper[4744]: E0106 16:23:44.964309 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:23:45 crc kubenswrapper[4744]: I0106 16:23:45.987790 4744 generic.go:334] "Generic (PLEG): container finished" podID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerID="7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44" exitCode=0 Jan 06 16:23:45 crc kubenswrapper[4744]: I0106 16:23:45.987906 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxk5l" event={"ID":"25f3e669-4a9e-48cd-be1c-50e76ecdef1f","Type":"ContainerDied","Data":"7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44"} Jan 06 16:23:47 crc kubenswrapper[4744]: I0106 16:23:47.006447 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxk5l" event={"ID":"25f3e669-4a9e-48cd-be1c-50e76ecdef1f","Type":"ContainerStarted","Data":"3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae"} Jan 06 16:23:47 crc kubenswrapper[4744]: I0106 16:23:47.030796 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fxk5l" podStartSLOduration=2.187216919 podStartE2EDuration="36.030779367s" podCreationTimestamp="2026-01-06 16:23:11 +0000 UTC" firstStartedPulling="2026-01-06 16:23:12.595207791 +0000 UTC m=+6389.222674149" lastFinishedPulling="2026-01-06 16:23:46.438770279 +0000 UTC m=+6423.066236597" observedRunningTime="2026-01-06 16:23:47.030279433 +0000 UTC m=+6423.657745751" watchObservedRunningTime="2026-01-06 16:23:47.030779367 +0000 UTC m=+6423.658245685" Jan 06 16:23:48 crc kubenswrapper[4744]: I0106 16:23:48.981132 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-fcpmw_6488ead4-081d-4faa-b3f3-f76b713b4aa9/nmstate-metrics/0.log" Jan 06 16:23:49 crc kubenswrapper[4744]: I0106 16:23:49.186421 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gkbrk_e89278c2-e058-4ea3-823f-00438d7ece52/nmstate-handler/0.log" Jan 06 16:23:49 crc kubenswrapper[4744]: I0106 16:23:49.188627 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-9l9t8_13ffe872-9861-485f-95a0-deaf5d02c62f/nmstate-console-plugin/0.log" Jan 06 16:23:49 crc kubenswrapper[4744]: I0106 16:23:49.192330 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-fcpmw_6488ead4-081d-4faa-b3f3-f76b713b4aa9/kube-rbac-proxy/0.log" Jan 06 16:23:49 crc kubenswrapper[4744]: I0106 16:23:49.903619 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-v4x2q_ac4dd1c4-7fdb-45a3-8727-e2ee45d3dc0a/nmstate-webhook/0.log" Jan 06 16:23:49 crc kubenswrapper[4744]: I0106 16:23:49.905773 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-d8hmc_c6660ee8-e4a2-4898-b8fc-ea4354bfc879/nmstate-operator/0.log" Jan 06 16:23:51 crc kubenswrapper[4744]: I0106 16:23:51.700143 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:51 crc kubenswrapper[4744]: I0106 16:23:51.700579 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:51 crc kubenswrapper[4744]: I0106 16:23:51.755044 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:52 crc kubenswrapper[4744]: I0106 16:23:52.127027 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:52 crc kubenswrapper[4744]: I0106 16:23:52.184545 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fxk5l"] Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.079894 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fxk5l" podUID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerName="registry-server" containerID="cri-o://3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae" gracePeriod=2 Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.585359 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.711435 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-catalog-content\") pod \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.711519 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-utilities\") pod \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.711695 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h6sb\" (UniqueName: \"kubernetes.io/projected/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-kube-api-access-7h6sb\") pod \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\" (UID: \"25f3e669-4a9e-48cd-be1c-50e76ecdef1f\") " Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.712146 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-utilities" (OuterVolumeSpecName: "utilities") pod "25f3e669-4a9e-48cd-be1c-50e76ecdef1f" (UID: "25f3e669-4a9e-48cd-be1c-50e76ecdef1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.713044 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.718343 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-kube-api-access-7h6sb" (OuterVolumeSpecName: "kube-api-access-7h6sb") pod "25f3e669-4a9e-48cd-be1c-50e76ecdef1f" (UID: "25f3e669-4a9e-48cd-be1c-50e76ecdef1f"). InnerVolumeSpecName "kube-api-access-7h6sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.756088 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25f3e669-4a9e-48cd-be1c-50e76ecdef1f" (UID: "25f3e669-4a9e-48cd-be1c-50e76ecdef1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.815606 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h6sb\" (UniqueName: \"kubernetes.io/projected/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-kube-api-access-7h6sb\") on node \"crc\" DevicePath \"\"" Jan 06 16:23:54 crc kubenswrapper[4744]: I0106 16:23:54.816210 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25f3e669-4a9e-48cd-be1c-50e76ecdef1f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.093206 4744 generic.go:334] "Generic (PLEG): container finished" podID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerID="3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae" exitCode=0 Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.093262 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxk5l" event={"ID":"25f3e669-4a9e-48cd-be1c-50e76ecdef1f","Type":"ContainerDied","Data":"3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae"} Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.093310 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxk5l" event={"ID":"25f3e669-4a9e-48cd-be1c-50e76ecdef1f","Type":"ContainerDied","Data":"ded4179d7afcd0d24c9f3d90abcb7b7180a2f8ac78d42f98eceaf962f0dca51e"} Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.093328 4744 scope.go:117] "RemoveContainer" containerID="3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.093339 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxk5l" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.119179 4744 scope.go:117] "RemoveContainer" containerID="7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.147068 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fxk5l"] Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.156620 4744 scope.go:117] "RemoveContainer" containerID="c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.158093 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fxk5l"] Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.195456 4744 scope.go:117] "RemoveContainer" containerID="3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae" Jan 06 16:23:55 crc kubenswrapper[4744]: E0106 16:23:55.195906 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae\": container with ID starting with 3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae not found: ID does not exist" containerID="3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.195963 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae"} err="failed to get container status \"3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae\": rpc error: code = NotFound desc = could not find container \"3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae\": container with ID starting with 3f1e9bdbb99b0381b86f8ee649d8259443e1d977a9b10e7d038d8ce614a59cae not found: ID does not exist" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.196005 4744 scope.go:117] "RemoveContainer" containerID="7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44" Jan 06 16:23:55 crc kubenswrapper[4744]: E0106 16:23:55.196518 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44\": container with ID starting with 7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44 not found: ID does not exist" containerID="7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.196568 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44"} err="failed to get container status \"7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44\": rpc error: code = NotFound desc = could not find container \"7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44\": container with ID starting with 7f8be0be69fabb4282e2cea4697f5aa6bd46fe9fca079c9042da87d092728c44 not found: ID does not exist" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.196602 4744 scope.go:117] "RemoveContainer" containerID="c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c" Jan 06 16:23:55 crc kubenswrapper[4744]: E0106 16:23:55.197131 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c\": container with ID starting with c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c not found: ID does not exist" containerID="c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.197237 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c"} err="failed to get container status \"c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c\": rpc error: code = NotFound desc = could not find container \"c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c\": container with ID starting with c1dc8be7a913036af34bfea50d0f5a32cfbdeaf9cea4c1bf2b69fcd334a5880c not found: ID does not exist" Jan 06 16:23:55 crc kubenswrapper[4744]: E0106 16:23:55.229840 4744 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25f3e669_4a9e_48cd_be1c_50e76ecdef1f.slice\": RecentStats: unable to find data in memory cache]" Jan 06 16:23:55 crc kubenswrapper[4744]: I0106 16:23:55.735340 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" path="/var/lib/kubelet/pods/25f3e669-4a9e-48cd-be1c-50e76ecdef1f/volumes" Jan 06 16:23:56 crc kubenswrapper[4744]: I0106 16:23:56.712044 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:23:56 crc kubenswrapper[4744]: E0106 16:23:56.712623 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:24:02 crc kubenswrapper[4744]: I0106 16:24:02.769598 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55d9dcdcb9-txlgb_34926775-1ace-452c-ae93-f25253e4b811/kube-rbac-proxy/0.log" Jan 06 16:24:02 crc kubenswrapper[4744]: I0106 16:24:02.820050 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55d9dcdcb9-txlgb_34926775-1ace-452c-ae93-f25253e4b811/manager/0.log" Jan 06 16:24:11 crc kubenswrapper[4744]: I0106 16:24:11.711116 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:24:11 crc kubenswrapper[4744]: E0106 16:24:11.711921 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:24:18 crc kubenswrapper[4744]: I0106 16:24:18.861605 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-79cf69ddc8-4qxck_8a59d77a-d5d2-4a5c-a63d-cdacacf67f0b/cluster-logging-operator/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.013555 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-95ctr_636a1772-2f55-430d-a282-35d995eeb134/collector/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.065581 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_a48ae2c1-a7b3-43e2-b7de-f2c6cafea7b9/loki-compactor/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.260314 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-5f678c8dd6-jtrh9_7549f443-f7d9-42fe-94a0-5bc9565caa14/loki-distributor/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.296098 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-b67674996-jsskr_ddffe4cb-6ab7-4556-aa91-620d374028db/gateway/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.304348 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-b67674996-jsskr_ddffe4cb-6ab7-4556-aa91-620d374028db/opa/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.461035 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-b67674996-w4jmm_4e939cdc-a027-4c89-b0e8-f5c0dff2af63/gateway/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.482239 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-b67674996-w4jmm_4e939cdc-a027-4c89-b0e8-f5c0dff2af63/opa/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.658910 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_d01a1f2a-5b58-4761-956c-199f1b7560fe/loki-index-gateway/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.753392 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_631bc775-d5be-445e-abd6-d6ec8afd9cf2/loki-ingester/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.907490 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-76788598db-v9bbz_a2c75046-2bab-4e58-b54a-0172540755a2/loki-querier/0.log" Jan 06 16:24:19 crc kubenswrapper[4744]: I0106 16:24:19.942602 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-69d9546745-xbtdk_e2a5e0ac-871e-456a-9fcb-04b991a7f6d6/loki-query-frontend/0.log" Jan 06 16:24:26 crc kubenswrapper[4744]: I0106 16:24:26.711188 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:24:26 crc kubenswrapper[4744]: E0106 16:24:26.712153 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:24:33 crc kubenswrapper[4744]: I0106 16:24:33.639532 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-wl5kh_19bdb736-8dea-45d9-b5cf-f9ee3407713a/kube-rbac-proxy/0.log" Jan 06 16:24:33 crc kubenswrapper[4744]: I0106 16:24:33.863638 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-wl5kh_19bdb736-8dea-45d9-b5cf-f9ee3407713a/controller/0.log" Jan 06 16:24:33 crc kubenswrapper[4744]: I0106 16:24:33.954634 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-frr-files/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.031975 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-reloader/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.046270 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-frr-files/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.070355 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-metrics/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.150855 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-reloader/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.326888 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-metrics/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.329653 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-reloader/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.341728 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-frr-files/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.365079 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-metrics/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.519133 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-frr-files/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.546510 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-metrics/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.565554 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/cp-reloader/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.606897 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/controller/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.770462 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/frr-metrics/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.781398 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/kube-rbac-proxy/0.log" Jan 06 16:24:34 crc kubenswrapper[4744]: I0106 16:24:34.841499 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/kube-rbac-proxy-frr/0.log" Jan 06 16:24:35 crc kubenswrapper[4744]: I0106 16:24:35.010371 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/reloader/0.log" Jan 06 16:24:35 crc kubenswrapper[4744]: I0106 16:24:35.087035 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-7t795_f49af91f-01db-4e5f-b442-11f20ec7a829/frr-k8s-webhook-server/0.log" Jan 06 16:24:35 crc kubenswrapper[4744]: I0106 16:24:35.311856 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-744888559c-nspzf_aadee3b1-2fa0-497f-b0cf-6e84bbf73d52/manager/0.log" Jan 06 16:24:35 crc kubenswrapper[4744]: I0106 16:24:35.429409 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-74784fd878-9zcxz_c4447713-2058-42f3-9b7f-0225267625e3/webhook-server/0.log" Jan 06 16:24:35 crc kubenswrapper[4744]: I0106 16:24:35.535942 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qzjdj_27d52d30-370e-44a9-82b1-eb0d7e197220/kube-rbac-proxy/0.log" Jan 06 16:24:36 crc kubenswrapper[4744]: I0106 16:24:36.368934 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qzjdj_27d52d30-370e-44a9-82b1-eb0d7e197220/speaker/0.log" Jan 06 16:24:36 crc kubenswrapper[4744]: I0106 16:24:36.629298 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mmvx7_4d920867-2ba6-42f6-acc3-266cbfbc7716/frr/0.log" Jan 06 16:24:37 crc kubenswrapper[4744]: I0106 16:24:37.711727 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:24:37 crc kubenswrapper[4744]: E0106 16:24:37.712334 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:24:48 crc kubenswrapper[4744]: I0106 16:24:48.711450 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:24:48 crc kubenswrapper[4744]: E0106 16:24:48.712244 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.174996 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6_69e01720-c541-4435-8ea7-3012016b80ff/util/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.331115 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6_69e01720-c541-4435-8ea7-3012016b80ff/util/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.369533 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6_69e01720-c541-4435-8ea7-3012016b80ff/pull/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.388205 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6_69e01720-c541-4435-8ea7-3012016b80ff/pull/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.547603 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6_69e01720-c541-4435-8ea7-3012016b80ff/util/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.595847 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6_69e01720-c541-4435-8ea7-3012016b80ff/pull/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.615838 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a2v9sp6_69e01720-c541-4435-8ea7-3012016b80ff/extract/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.747770 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2_9fc7c9b1-f812-4104-8a87-b9d3024fc25e/util/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.914994 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2_9fc7c9b1-f812-4104-8a87-b9d3024fc25e/util/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.922686 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2_9fc7c9b1-f812-4104-8a87-b9d3024fc25e/pull/0.log" Jan 06 16:24:49 crc kubenswrapper[4744]: I0106 16:24:49.941496 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2_9fc7c9b1-f812-4104-8a87-b9d3024fc25e/pull/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.174063 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2_9fc7c9b1-f812-4104-8a87-b9d3024fc25e/pull/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.179733 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2_9fc7c9b1-f812-4104-8a87-b9d3024fc25e/util/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.182985 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360b7rkf2_9fc7c9b1-f812-4104-8a87-b9d3024fc25e/extract/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.349063 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt_e1fc4c9c-6c58-420a-9132-1c9926cd00eb/util/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.533685 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt_e1fc4c9c-6c58-420a-9132-1c9926cd00eb/pull/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.551859 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt_e1fc4c9c-6c58-420a-9132-1c9926cd00eb/pull/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.560738 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt_e1fc4c9c-6c58-420a-9132-1c9926cd00eb/util/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.714344 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt_e1fc4c9c-6c58-420a-9132-1c9926cd00eb/pull/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.720838 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt_e1fc4c9c-6c58-420a-9132-1c9926cd00eb/util/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.762047 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d49wjzt_e1fc4c9c-6c58-420a-9132-1c9926cd00eb/extract/0.log" Jan 06 16:24:50 crc kubenswrapper[4744]: I0106 16:24:50.908897 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2_7b10c0ca-7d3b-436c-be4e-80798af3e660/util/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.102497 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2_7b10c0ca-7d3b-436c-be4e-80798af3e660/pull/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.103745 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2_7b10c0ca-7d3b-436c-be4e-80798af3e660/util/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.148853 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2_7b10c0ca-7d3b-436c-be4e-80798af3e660/pull/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.306484 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2_7b10c0ca-7d3b-436c-be4e-80798af3e660/util/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.310697 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2_7b10c0ca-7d3b-436c-be4e-80798af3e660/pull/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.360204 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8lzmh2_7b10c0ca-7d3b-436c-be4e-80798af3e660/extract/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.515744 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f_e99516eb-a6a2-4cf2-b9cd-5f67cf02905d/util/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.663395 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f_e99516eb-a6a2-4cf2-b9cd-5f67cf02905d/util/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.669997 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f_e99516eb-a6a2-4cf2-b9cd-5f67cf02905d/pull/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.708605 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f_e99516eb-a6a2-4cf2-b9cd-5f67cf02905d/pull/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.926414 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f_e99516eb-a6a2-4cf2-b9cd-5f67cf02905d/util/0.log" Jan 06 16:24:51 crc kubenswrapper[4744]: I0106 16:24:51.989873 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f_e99516eb-a6a2-4cf2-b9cd-5f67cf02905d/pull/0.log" Jan 06 16:24:52 crc kubenswrapper[4744]: I0106 16:24:52.008738 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zc22f_e99516eb-a6a2-4cf2-b9cd-5f67cf02905d/extract/0.log" Jan 06 16:24:52 crc kubenswrapper[4744]: I0106 16:24:52.171922 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sqlqh_bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df/extract-utilities/0.log" Jan 06 16:24:52 crc kubenswrapper[4744]: I0106 16:24:52.385964 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sqlqh_bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df/extract-utilities/0.log" Jan 06 16:24:52 crc kubenswrapper[4744]: I0106 16:24:52.471429 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sqlqh_bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df/extract-content/0.log" Jan 06 16:24:52 crc kubenswrapper[4744]: I0106 16:24:52.492881 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sqlqh_bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df/extract-content/0.log" Jan 06 16:24:52 crc kubenswrapper[4744]: I0106 16:24:52.625895 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sqlqh_bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df/extract-utilities/0.log" Jan 06 16:24:52 crc kubenswrapper[4744]: I0106 16:24:52.654465 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sqlqh_bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df/extract-content/0.log" Jan 06 16:24:52 crc kubenswrapper[4744]: I0106 16:24:52.753856 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8755w_889a8fc3-2966-446c-96d1-fa9203b19280/extract-utilities/0.log" Jan 06 16:24:53 crc kubenswrapper[4744]: I0106 16:24:53.076624 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8755w_889a8fc3-2966-446c-96d1-fa9203b19280/extract-content/0.log" Jan 06 16:24:53 crc kubenswrapper[4744]: I0106 16:24:53.148401 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8755w_889a8fc3-2966-446c-96d1-fa9203b19280/extract-utilities/0.log" Jan 06 16:24:53 crc kubenswrapper[4744]: I0106 16:24:53.162319 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8755w_889a8fc3-2966-446c-96d1-fa9203b19280/extract-content/0.log" Jan 06 16:24:53 crc kubenswrapper[4744]: I0106 16:24:53.481513 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-sqlqh_bbeaa4ea-0334-48cd-9ddd-f7a5ea6ac2df/registry-server/0.log" Jan 06 16:24:53 crc kubenswrapper[4744]: I0106 16:24:53.586692 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8755w_889a8fc3-2966-446c-96d1-fa9203b19280/extract-utilities/0.log" Jan 06 16:24:53 crc kubenswrapper[4744]: I0106 16:24:53.680622 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8755w_889a8fc3-2966-446c-96d1-fa9203b19280/extract-content/0.log" Jan 06 16:24:53 crc kubenswrapper[4744]: I0106 16:24:53.700213 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-wsc5g_e6c65897-076d-4e90-a28a-2989a57e11be/marketplace-operator/0.log" Jan 06 16:24:53 crc kubenswrapper[4744]: I0106 16:24:53.868227 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pssbh_d470584a-0f48-436b-9ead-06f8947b7115/extract-utilities/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.037337 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pssbh_d470584a-0f48-436b-9ead-06f8947b7115/extract-utilities/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.099846 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pssbh_d470584a-0f48-436b-9ead-06f8947b7115/extract-content/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.178509 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pssbh_d470584a-0f48-436b-9ead-06f8947b7115/extract-content/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.355283 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pssbh_d470584a-0f48-436b-9ead-06f8947b7115/extract-utilities/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.410671 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pssbh_d470584a-0f48-436b-9ead-06f8947b7115/extract-content/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.584808 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrxvs_78579813-4567-46c4-8353-4e6418ba7722/extract-utilities/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.716699 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pssbh_d470584a-0f48-436b-9ead-06f8947b7115/registry-server/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.789715 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrxvs_78579813-4567-46c4-8353-4e6418ba7722/extract-content/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.793028 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrxvs_78579813-4567-46c4-8353-4e6418ba7722/extract-utilities/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.804938 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrxvs_78579813-4567-46c4-8353-4e6418ba7722/extract-content/0.log" Jan 06 16:24:54 crc kubenswrapper[4744]: I0106 16:24:54.919431 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-8755w_889a8fc3-2966-446c-96d1-fa9203b19280/registry-server/0.log" Jan 06 16:24:55 crc kubenswrapper[4744]: I0106 16:24:55.049784 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrxvs_78579813-4567-46c4-8353-4e6418ba7722/extract-utilities/0.log" Jan 06 16:24:55 crc kubenswrapper[4744]: I0106 16:24:55.070860 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrxvs_78579813-4567-46c4-8353-4e6418ba7722/extract-content/0.log" Jan 06 16:24:55 crc kubenswrapper[4744]: I0106 16:24:55.957936 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vrxvs_78579813-4567-46c4-8353-4e6418ba7722/registry-server/0.log" Jan 06 16:24:59 crc kubenswrapper[4744]: I0106 16:24:59.711540 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:24:59 crc kubenswrapper[4744]: E0106 16:24:59.712538 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.188602 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5pcxw"] Jan 06 16:25:07 crc kubenswrapper[4744]: E0106 16:25:07.189657 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerName="registry-server" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.189675 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerName="registry-server" Jan 06 16:25:07 crc kubenswrapper[4744]: E0106 16:25:07.189686 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerName="extract-content" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.189695 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerName="extract-content" Jan 06 16:25:07 crc kubenswrapper[4744]: E0106 16:25:07.189729 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerName="extract-utilities" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.189738 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerName="extract-utilities" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.190025 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="25f3e669-4a9e-48cd-be1c-50e76ecdef1f" containerName="registry-server" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.192017 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.205898 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pcxw"] Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.269561 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-utilities\") pod \"redhat-marketplace-5pcxw\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.269718 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-catalog-content\") pod \"redhat-marketplace-5pcxw\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.269811 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzp88\" (UniqueName: \"kubernetes.io/projected/5450550a-95c9-434c-9d0d-36767231d6f4-kube-api-access-vzp88\") pod \"redhat-marketplace-5pcxw\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.371842 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-catalog-content\") pod \"redhat-marketplace-5pcxw\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.371975 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzp88\" (UniqueName: \"kubernetes.io/projected/5450550a-95c9-434c-9d0d-36767231d6f4-kube-api-access-vzp88\") pod \"redhat-marketplace-5pcxw\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.372025 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-utilities\") pod \"redhat-marketplace-5pcxw\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.372475 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-utilities\") pod \"redhat-marketplace-5pcxw\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.372484 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-catalog-content\") pod \"redhat-marketplace-5pcxw\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.392683 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzp88\" (UniqueName: \"kubernetes.io/projected/5450550a-95c9-434c-9d0d-36767231d6f4-kube-api-access-vzp88\") pod \"redhat-marketplace-5pcxw\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:07 crc kubenswrapper[4744]: I0106 16:25:07.516908 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:08 crc kubenswrapper[4744]: I0106 16:25:08.061383 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pcxw"] Jan 06 16:25:08 crc kubenswrapper[4744]: I0106 16:25:08.975910 4744 generic.go:334] "Generic (PLEG): container finished" podID="5450550a-95c9-434c-9d0d-36767231d6f4" containerID="42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9" exitCode=0 Jan 06 16:25:08 crc kubenswrapper[4744]: I0106 16:25:08.975973 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pcxw" event={"ID":"5450550a-95c9-434c-9d0d-36767231d6f4","Type":"ContainerDied","Data":"42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9"} Jan 06 16:25:08 crc kubenswrapper[4744]: I0106 16:25:08.976228 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pcxw" event={"ID":"5450550a-95c9-434c-9d0d-36767231d6f4","Type":"ContainerStarted","Data":"4f68c8d65371bedc773a6863c7a76aaa5d90345ad025816b6a8d71477b4308f1"} Jan 06 16:25:09 crc kubenswrapper[4744]: I0106 16:25:09.508749 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-q677r_7f4652da-367f-462e-92cb-cd4a7acc9487/prometheus-operator/0.log" Jan 06 16:25:09 crc kubenswrapper[4744]: I0106 16:25:09.694214 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-bb7f9df74-zmbfj_6254c369-a75d-410f-8dc8-1ab47bab3908/prometheus-operator-admission-webhook/0.log" Jan 06 16:25:09 crc kubenswrapper[4744]: I0106 16:25:09.714328 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-bb7f9df74-rstzz_e85ae009-0426-4c0e-a1af-de7cd66148cb/prometheus-operator-admission-webhook/0.log" Jan 06 16:25:09 crc kubenswrapper[4744]: I0106 16:25:09.914549 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-66cbf594b5-mbspt_b3368f37-f50b-431f-9900-c17002a24788/observability-ui-dashboards/0.log" Jan 06 16:25:09 crc kubenswrapper[4744]: I0106 16:25:09.924082 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-cklnq_313a5ded-3973-4f75-a2ff-d112d168de8d/operator/0.log" Jan 06 16:25:09 crc kubenswrapper[4744]: I0106 16:25:09.990205 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pcxw" event={"ID":"5450550a-95c9-434c-9d0d-36767231d6f4","Type":"ContainerStarted","Data":"749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd"} Jan 06 16:25:10 crc kubenswrapper[4744]: I0106 16:25:10.171908 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-rx8pm_81494dce-a93e-4e68-9276-765b36e2cde4/perses-operator/0.log" Jan 06 16:25:11 crc kubenswrapper[4744]: I0106 16:25:11.004136 4744 generic.go:334] "Generic (PLEG): container finished" podID="5450550a-95c9-434c-9d0d-36767231d6f4" containerID="749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd" exitCode=0 Jan 06 16:25:11 crc kubenswrapper[4744]: I0106 16:25:11.004212 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pcxw" event={"ID":"5450550a-95c9-434c-9d0d-36767231d6f4","Type":"ContainerDied","Data":"749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd"} Jan 06 16:25:12 crc kubenswrapper[4744]: I0106 16:25:12.711455 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:25:12 crc kubenswrapper[4744]: E0106 16:25:12.712278 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:25:13 crc kubenswrapper[4744]: I0106 16:25:13.028512 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pcxw" event={"ID":"5450550a-95c9-434c-9d0d-36767231d6f4","Type":"ContainerStarted","Data":"a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35"} Jan 06 16:25:13 crc kubenswrapper[4744]: I0106 16:25:13.050916 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5pcxw" podStartSLOduration=2.890690645 podStartE2EDuration="6.050898803s" podCreationTimestamp="2026-01-06 16:25:07 +0000 UTC" firstStartedPulling="2026-01-06 16:25:08.979012248 +0000 UTC m=+6505.606478566" lastFinishedPulling="2026-01-06 16:25:12.139220406 +0000 UTC m=+6508.766686724" observedRunningTime="2026-01-06 16:25:13.045778587 +0000 UTC m=+6509.673244905" watchObservedRunningTime="2026-01-06 16:25:13.050898803 +0000 UTC m=+6509.678365121" Jan 06 16:25:17 crc kubenswrapper[4744]: I0106 16:25:17.517229 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:17 crc kubenswrapper[4744]: I0106 16:25:17.518298 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:17 crc kubenswrapper[4744]: I0106 16:25:17.604974 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:18 crc kubenswrapper[4744]: I0106 16:25:18.126959 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:18 crc kubenswrapper[4744]: I0106 16:25:18.171705 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pcxw"] Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.102023 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5pcxw" podUID="5450550a-95c9-434c-9d0d-36767231d6f4" containerName="registry-server" containerID="cri-o://a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35" gracePeriod=2 Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.631503 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.809256 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzp88\" (UniqueName: \"kubernetes.io/projected/5450550a-95c9-434c-9d0d-36767231d6f4-kube-api-access-vzp88\") pod \"5450550a-95c9-434c-9d0d-36767231d6f4\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.809424 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-catalog-content\") pod \"5450550a-95c9-434c-9d0d-36767231d6f4\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.809660 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-utilities\") pod \"5450550a-95c9-434c-9d0d-36767231d6f4\" (UID: \"5450550a-95c9-434c-9d0d-36767231d6f4\") " Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.810252 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-utilities" (OuterVolumeSpecName: "utilities") pod "5450550a-95c9-434c-9d0d-36767231d6f4" (UID: "5450550a-95c9-434c-9d0d-36767231d6f4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.814921 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5450550a-95c9-434c-9d0d-36767231d6f4-kube-api-access-vzp88" (OuterVolumeSpecName: "kube-api-access-vzp88") pod "5450550a-95c9-434c-9d0d-36767231d6f4" (UID: "5450550a-95c9-434c-9d0d-36767231d6f4"). InnerVolumeSpecName "kube-api-access-vzp88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.830219 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5450550a-95c9-434c-9d0d-36767231d6f4" (UID: "5450550a-95c9-434c-9d0d-36767231d6f4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.913147 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.913208 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzp88\" (UniqueName: \"kubernetes.io/projected/5450550a-95c9-434c-9d0d-36767231d6f4-kube-api-access-vzp88\") on node \"crc\" DevicePath \"\"" Jan 06 16:25:20 crc kubenswrapper[4744]: I0106 16:25:20.913227 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5450550a-95c9-434c-9d0d-36767231d6f4-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.118267 4744 generic.go:334] "Generic (PLEG): container finished" podID="5450550a-95c9-434c-9d0d-36767231d6f4" containerID="a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35" exitCode=0 Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.118397 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pcxw" event={"ID":"5450550a-95c9-434c-9d0d-36767231d6f4","Type":"ContainerDied","Data":"a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35"} Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.119331 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5pcxw" event={"ID":"5450550a-95c9-434c-9d0d-36767231d6f4","Type":"ContainerDied","Data":"4f68c8d65371bedc773a6863c7a76aaa5d90345ad025816b6a8d71477b4308f1"} Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.119372 4744 scope.go:117] "RemoveContainer" containerID="a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.118453 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5pcxw" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.147962 4744 scope.go:117] "RemoveContainer" containerID="749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.166758 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pcxw"] Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.181715 4744 scope.go:117] "RemoveContainer" containerID="42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.187140 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5pcxw"] Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.252617 4744 scope.go:117] "RemoveContainer" containerID="a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35" Jan 06 16:25:21 crc kubenswrapper[4744]: E0106 16:25:21.253148 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35\": container with ID starting with a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35 not found: ID does not exist" containerID="a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.253190 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35"} err="failed to get container status \"a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35\": rpc error: code = NotFound desc = could not find container \"a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35\": container with ID starting with a8457910cab30994f38d188423a4c5c5c287e9a7b25b1ddc92e911e91234ee35 not found: ID does not exist" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.253211 4744 scope.go:117] "RemoveContainer" containerID="749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd" Jan 06 16:25:21 crc kubenswrapper[4744]: E0106 16:25:21.253545 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd\": container with ID starting with 749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd not found: ID does not exist" containerID="749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.253570 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd"} err="failed to get container status \"749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd\": rpc error: code = NotFound desc = could not find container \"749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd\": container with ID starting with 749edf015db9d95ee99ae9f9e48ab5052d915c73146d356eb17d3771c798abfd not found: ID does not exist" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.253590 4744 scope.go:117] "RemoveContainer" containerID="42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9" Jan 06 16:25:21 crc kubenswrapper[4744]: E0106 16:25:21.253935 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9\": container with ID starting with 42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9 not found: ID does not exist" containerID="42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.253962 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9"} err="failed to get container status \"42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9\": rpc error: code = NotFound desc = could not find container \"42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9\": container with ID starting with 42f350472d1ed3da386ddc457ca9dbb65f1af28dab761b3a062d4d074695c1d9 not found: ID does not exist" Jan 06 16:25:21 crc kubenswrapper[4744]: I0106 16:25:21.725905 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5450550a-95c9-434c-9d0d-36767231d6f4" path="/var/lib/kubelet/pods/5450550a-95c9-434c-9d0d-36767231d6f4/volumes" Jan 06 16:25:23 crc kubenswrapper[4744]: I0106 16:25:23.718578 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:25:23 crc kubenswrapper[4744]: E0106 16:25:23.719138 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:25:24 crc kubenswrapper[4744]: I0106 16:25:24.514976 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55d9dcdcb9-txlgb_34926775-1ace-452c-ae93-f25253e4b811/kube-rbac-proxy/0.log" Jan 06 16:25:24 crc kubenswrapper[4744]: I0106 16:25:24.560810 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-55d9dcdcb9-txlgb_34926775-1ace-452c-ae93-f25253e4b811/manager/0.log" Jan 06 16:25:36 crc kubenswrapper[4744]: I0106 16:25:36.711515 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:25:36 crc kubenswrapper[4744]: E0106 16:25:36.712406 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:25:50 crc kubenswrapper[4744]: I0106 16:25:50.711296 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:25:50 crc kubenswrapper[4744]: E0106 16:25:50.712031 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:26:03 crc kubenswrapper[4744]: I0106 16:26:03.727831 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:26:03 crc kubenswrapper[4744]: E0106 16:26:03.729284 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:26:18 crc kubenswrapper[4744]: I0106 16:26:18.712738 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:26:18 crc kubenswrapper[4744]: E0106 16:26:18.714144 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:26:31 crc kubenswrapper[4744]: I0106 16:26:31.712197 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:26:31 crc kubenswrapper[4744]: E0106 16:26:31.712931 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:26:43 crc kubenswrapper[4744]: I0106 16:26:43.725944 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:26:43 crc kubenswrapper[4744]: E0106 16:26:43.727828 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:26:56 crc kubenswrapper[4744]: I0106 16:26:56.711261 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:26:56 crc kubenswrapper[4744]: E0106 16:26:56.713458 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:27:03 crc kubenswrapper[4744]: I0106 16:27:03.438565 4744 generic.go:334] "Generic (PLEG): container finished" podID="d643fddf-ee4c-4226-814c-4acc9a13eb91" containerID="bacda9e92884680457b14ff3b2d1ebdaea4498c05385abf6b2b2c9ef58a12109" exitCode=0 Jan 06 16:27:03 crc kubenswrapper[4744]: I0106 16:27:03.438658 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8vxws/must-gather-vqjqk" event={"ID":"d643fddf-ee4c-4226-814c-4acc9a13eb91","Type":"ContainerDied","Data":"bacda9e92884680457b14ff3b2d1ebdaea4498c05385abf6b2b2c9ef58a12109"} Jan 06 16:27:03 crc kubenswrapper[4744]: I0106 16:27:03.440341 4744 scope.go:117] "RemoveContainer" containerID="bacda9e92884680457b14ff3b2d1ebdaea4498c05385abf6b2b2c9ef58a12109" Jan 06 16:27:04 crc kubenswrapper[4744]: I0106 16:27:04.338113 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8vxws_must-gather-vqjqk_d643fddf-ee4c-4226-814c-4acc9a13eb91/gather/0.log" Jan 06 16:27:10 crc kubenswrapper[4744]: I0106 16:27:10.711804 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:27:10 crc kubenswrapper[4744]: E0106 16:27:10.712778 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.019133 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8vxws/must-gather-vqjqk"] Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.019621 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-8vxws/must-gather-vqjqk" podUID="d643fddf-ee4c-4226-814c-4acc9a13eb91" containerName="copy" containerID="cri-o://4576b7e4ded6902d80bb24939deab06f7cef9aa30bf17f8bbe287a3819e73fe9" gracePeriod=2 Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.034463 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8vxws/must-gather-vqjqk"] Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.553065 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8vxws_must-gather-vqjqk_d643fddf-ee4c-4226-814c-4acc9a13eb91/copy/0.log" Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.553868 4744 generic.go:334] "Generic (PLEG): container finished" podID="d643fddf-ee4c-4226-814c-4acc9a13eb91" containerID="4576b7e4ded6902d80bb24939deab06f7cef9aa30bf17f8bbe287a3819e73fe9" exitCode=143 Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.669570 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8vxws_must-gather-vqjqk_d643fddf-ee4c-4226-814c-4acc9a13eb91/copy/0.log" Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.670330 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.837970 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqjqv\" (UniqueName: \"kubernetes.io/projected/d643fddf-ee4c-4226-814c-4acc9a13eb91-kube-api-access-lqjqv\") pod \"d643fddf-ee4c-4226-814c-4acc9a13eb91\" (UID: \"d643fddf-ee4c-4226-814c-4acc9a13eb91\") " Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.838040 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d643fddf-ee4c-4226-814c-4acc9a13eb91-must-gather-output\") pod \"d643fddf-ee4c-4226-814c-4acc9a13eb91\" (UID: \"d643fddf-ee4c-4226-814c-4acc9a13eb91\") " Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.845735 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d643fddf-ee4c-4226-814c-4acc9a13eb91-kube-api-access-lqjqv" (OuterVolumeSpecName: "kube-api-access-lqjqv") pod "d643fddf-ee4c-4226-814c-4acc9a13eb91" (UID: "d643fddf-ee4c-4226-814c-4acc9a13eb91"). InnerVolumeSpecName "kube-api-access-lqjqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:27:12 crc kubenswrapper[4744]: I0106 16:27:12.941582 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqjqv\" (UniqueName: \"kubernetes.io/projected/d643fddf-ee4c-4226-814c-4acc9a13eb91-kube-api-access-lqjqv\") on node \"crc\" DevicePath \"\"" Jan 06 16:27:13 crc kubenswrapper[4744]: I0106 16:27:13.013016 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d643fddf-ee4c-4226-814c-4acc9a13eb91-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d643fddf-ee4c-4226-814c-4acc9a13eb91" (UID: "d643fddf-ee4c-4226-814c-4acc9a13eb91"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:27:13 crc kubenswrapper[4744]: I0106 16:27:13.044439 4744 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d643fddf-ee4c-4226-814c-4acc9a13eb91-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 06 16:27:13 crc kubenswrapper[4744]: I0106 16:27:13.568346 4744 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8vxws_must-gather-vqjqk_d643fddf-ee4c-4226-814c-4acc9a13eb91/copy/0.log" Jan 06 16:27:13 crc kubenswrapper[4744]: I0106 16:27:13.569439 4744 scope.go:117] "RemoveContainer" containerID="4576b7e4ded6902d80bb24939deab06f7cef9aa30bf17f8bbe287a3819e73fe9" Jan 06 16:27:13 crc kubenswrapper[4744]: I0106 16:27:13.569495 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8vxws/must-gather-vqjqk" Jan 06 16:27:13 crc kubenswrapper[4744]: I0106 16:27:13.590983 4744 scope.go:117] "RemoveContainer" containerID="bacda9e92884680457b14ff3b2d1ebdaea4498c05385abf6b2b2c9ef58a12109" Jan 06 16:27:13 crc kubenswrapper[4744]: I0106 16:27:13.727433 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d643fddf-ee4c-4226-814c-4acc9a13eb91" path="/var/lib/kubelet/pods/d643fddf-ee4c-4226-814c-4acc9a13eb91/volumes" Jan 06 16:27:14 crc kubenswrapper[4744]: I0106 16:27:14.969631 4744 scope.go:117] "RemoveContainer" containerID="949c3e1bbd5fda367b0781011a622105ea2709def5c7db93597c921229a850f2" Jan 06 16:27:23 crc kubenswrapper[4744]: I0106 16:27:23.725531 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:27:23 crc kubenswrapper[4744]: E0106 16:27:23.726756 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:27:35 crc kubenswrapper[4744]: I0106 16:27:35.711669 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:27:35 crc kubenswrapper[4744]: E0106 16:27:35.712650 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:27:46 crc kubenswrapper[4744]: I0106 16:27:46.712284 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:27:46 crc kubenswrapper[4744]: E0106 16:27:46.713340 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:27:57 crc kubenswrapper[4744]: I0106 16:27:57.711414 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:27:57 crc kubenswrapper[4744]: E0106 16:27:57.712296 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.516617 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pwrx9"] Jan 06 16:27:58 crc kubenswrapper[4744]: E0106 16:27:58.517535 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5450550a-95c9-434c-9d0d-36767231d6f4" containerName="extract-content" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.517555 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5450550a-95c9-434c-9d0d-36767231d6f4" containerName="extract-content" Jan 06 16:27:58 crc kubenswrapper[4744]: E0106 16:27:58.517595 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5450550a-95c9-434c-9d0d-36767231d6f4" containerName="registry-server" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.517603 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5450550a-95c9-434c-9d0d-36767231d6f4" containerName="registry-server" Jan 06 16:27:58 crc kubenswrapper[4744]: E0106 16:27:58.517616 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d643fddf-ee4c-4226-814c-4acc9a13eb91" containerName="gather" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.517623 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d643fddf-ee4c-4226-814c-4acc9a13eb91" containerName="gather" Jan 06 16:27:58 crc kubenswrapper[4744]: E0106 16:27:58.517644 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d643fddf-ee4c-4226-814c-4acc9a13eb91" containerName="copy" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.517651 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="d643fddf-ee4c-4226-814c-4acc9a13eb91" containerName="copy" Jan 06 16:27:58 crc kubenswrapper[4744]: E0106 16:27:58.517667 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5450550a-95c9-434c-9d0d-36767231d6f4" containerName="extract-utilities" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.517675 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="5450550a-95c9-434c-9d0d-36767231d6f4" containerName="extract-utilities" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.517934 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d643fddf-ee4c-4226-814c-4acc9a13eb91" containerName="gather" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.517955 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="5450550a-95c9-434c-9d0d-36767231d6f4" containerName="registry-server" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.517978 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="d643fddf-ee4c-4226-814c-4acc9a13eb91" containerName="copy" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.519780 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.532678 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pwrx9"] Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.597149 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-catalog-content\") pod \"community-operators-pwrx9\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.597319 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h47h\" (UniqueName: \"kubernetes.io/projected/aae9105d-b113-4d86-9f24-733246388a8e-kube-api-access-7h47h\") pod \"community-operators-pwrx9\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.597397 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-utilities\") pod \"community-operators-pwrx9\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.699719 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h47h\" (UniqueName: \"kubernetes.io/projected/aae9105d-b113-4d86-9f24-733246388a8e-kube-api-access-7h47h\") pod \"community-operators-pwrx9\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.700066 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-utilities\") pod \"community-operators-pwrx9\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.700210 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-catalog-content\") pod \"community-operators-pwrx9\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.700522 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-utilities\") pod \"community-operators-pwrx9\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.700794 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-catalog-content\") pod \"community-operators-pwrx9\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.719930 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h47h\" (UniqueName: \"kubernetes.io/projected/aae9105d-b113-4d86-9f24-733246388a8e-kube-api-access-7h47h\") pod \"community-operators-pwrx9\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:58 crc kubenswrapper[4744]: I0106 16:27:58.853851 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:27:59 crc kubenswrapper[4744]: I0106 16:27:59.378139 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pwrx9"] Jan 06 16:28:00 crc kubenswrapper[4744]: I0106 16:28:00.155912 4744 generic.go:334] "Generic (PLEG): container finished" podID="aae9105d-b113-4d86-9f24-733246388a8e" containerID="bf39b04c1a23ca5caa6997ff99c5e11325584e62f2d6fb39f08316656cc07b7d" exitCode=0 Jan 06 16:28:00 crc kubenswrapper[4744]: I0106 16:28:00.155994 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwrx9" event={"ID":"aae9105d-b113-4d86-9f24-733246388a8e","Type":"ContainerDied","Data":"bf39b04c1a23ca5caa6997ff99c5e11325584e62f2d6fb39f08316656cc07b7d"} Jan 06 16:28:00 crc kubenswrapper[4744]: I0106 16:28:00.156350 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwrx9" event={"ID":"aae9105d-b113-4d86-9f24-733246388a8e","Type":"ContainerStarted","Data":"8b626c371311a5bd2a19a4c61283ce85d84ec02b3c55e92b15e78ff3d01deb12"} Jan 06 16:28:02 crc kubenswrapper[4744]: I0106 16:28:02.182695 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwrx9" event={"ID":"aae9105d-b113-4d86-9f24-733246388a8e","Type":"ContainerStarted","Data":"2bdffd01556d231effa34eee304501d73dd353bab0d6dd9b4c8f1383f2737111"} Jan 06 16:28:03 crc kubenswrapper[4744]: I0106 16:28:03.197904 4744 generic.go:334] "Generic (PLEG): container finished" podID="aae9105d-b113-4d86-9f24-733246388a8e" containerID="2bdffd01556d231effa34eee304501d73dd353bab0d6dd9b4c8f1383f2737111" exitCode=0 Jan 06 16:28:03 crc kubenswrapper[4744]: I0106 16:28:03.198298 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwrx9" event={"ID":"aae9105d-b113-4d86-9f24-733246388a8e","Type":"ContainerDied","Data":"2bdffd01556d231effa34eee304501d73dd353bab0d6dd9b4c8f1383f2737111"} Jan 06 16:28:04 crc kubenswrapper[4744]: I0106 16:28:04.209643 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwrx9" event={"ID":"aae9105d-b113-4d86-9f24-733246388a8e","Type":"ContainerStarted","Data":"cf5747b6ae753a1eb30cb822a2550605b3b89dddc8ca3318f6c18da959f218f8"} Jan 06 16:28:04 crc kubenswrapper[4744]: I0106 16:28:04.229870 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pwrx9" podStartSLOduration=2.670350301 podStartE2EDuration="6.229853688s" podCreationTimestamp="2026-01-06 16:27:58 +0000 UTC" firstStartedPulling="2026-01-06 16:28:00.158818535 +0000 UTC m=+6676.786284883" lastFinishedPulling="2026-01-06 16:28:03.718321962 +0000 UTC m=+6680.345788270" observedRunningTime="2026-01-06 16:28:04.225750838 +0000 UTC m=+6680.853217166" watchObservedRunningTime="2026-01-06 16:28:04.229853688 +0000 UTC m=+6680.857319996" Jan 06 16:28:08 crc kubenswrapper[4744]: I0106 16:28:08.854131 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:28:08 crc kubenswrapper[4744]: I0106 16:28:08.854916 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:28:08 crc kubenswrapper[4744]: I0106 16:28:08.923930 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:28:09 crc kubenswrapper[4744]: I0106 16:28:09.335523 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:28:11 crc kubenswrapper[4744]: I0106 16:28:11.731436 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:28:11 crc kubenswrapper[4744]: E0106 16:28:11.735232 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:28:12 crc kubenswrapper[4744]: I0106 16:28:12.433576 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pwrx9"] Jan 06 16:28:12 crc kubenswrapper[4744]: I0106 16:28:12.433873 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pwrx9" podUID="aae9105d-b113-4d86-9f24-733246388a8e" containerName="registry-server" containerID="cri-o://cf5747b6ae753a1eb30cb822a2550605b3b89dddc8ca3318f6c18da959f218f8" gracePeriod=2 Jan 06 16:28:13 crc kubenswrapper[4744]: I0106 16:28:13.311339 4744 generic.go:334] "Generic (PLEG): container finished" podID="aae9105d-b113-4d86-9f24-733246388a8e" containerID="cf5747b6ae753a1eb30cb822a2550605b3b89dddc8ca3318f6c18da959f218f8" exitCode=0 Jan 06 16:28:13 crc kubenswrapper[4744]: I0106 16:28:13.311431 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwrx9" event={"ID":"aae9105d-b113-4d86-9f24-733246388a8e","Type":"ContainerDied","Data":"cf5747b6ae753a1eb30cb822a2550605b3b89dddc8ca3318f6c18da959f218f8"} Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.042093 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.176253 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-catalog-content\") pod \"aae9105d-b113-4d86-9f24-733246388a8e\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.176420 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h47h\" (UniqueName: \"kubernetes.io/projected/aae9105d-b113-4d86-9f24-733246388a8e-kube-api-access-7h47h\") pod \"aae9105d-b113-4d86-9f24-733246388a8e\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.176579 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-utilities\") pod \"aae9105d-b113-4d86-9f24-733246388a8e\" (UID: \"aae9105d-b113-4d86-9f24-733246388a8e\") " Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.178307 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-utilities" (OuterVolumeSpecName: "utilities") pod "aae9105d-b113-4d86-9f24-733246388a8e" (UID: "aae9105d-b113-4d86-9f24-733246388a8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.184546 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae9105d-b113-4d86-9f24-733246388a8e-kube-api-access-7h47h" (OuterVolumeSpecName: "kube-api-access-7h47h") pod "aae9105d-b113-4d86-9f24-733246388a8e" (UID: "aae9105d-b113-4d86-9f24-733246388a8e"). InnerVolumeSpecName "kube-api-access-7h47h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.226426 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aae9105d-b113-4d86-9f24-733246388a8e" (UID: "aae9105d-b113-4d86-9f24-733246388a8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.280203 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h47h\" (UniqueName: \"kubernetes.io/projected/aae9105d-b113-4d86-9f24-733246388a8e-kube-api-access-7h47h\") on node \"crc\" DevicePath \"\"" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.280228 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.280237 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aae9105d-b113-4d86-9f24-733246388a8e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.325624 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pwrx9" event={"ID":"aae9105d-b113-4d86-9f24-733246388a8e","Type":"ContainerDied","Data":"8b626c371311a5bd2a19a4c61283ce85d84ec02b3c55e92b15e78ff3d01deb12"} Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.325671 4744 scope.go:117] "RemoveContainer" containerID="cf5747b6ae753a1eb30cb822a2550605b3b89dddc8ca3318f6c18da959f218f8" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.325717 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pwrx9" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.360481 4744 scope.go:117] "RemoveContainer" containerID="2bdffd01556d231effa34eee304501d73dd353bab0d6dd9b4c8f1383f2737111" Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.377061 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pwrx9"] Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.392592 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pwrx9"] Jan 06 16:28:14 crc kubenswrapper[4744]: I0106 16:28:14.404238 4744 scope.go:117] "RemoveContainer" containerID="bf39b04c1a23ca5caa6997ff99c5e11325584e62f2d6fb39f08316656cc07b7d" Jan 06 16:28:15 crc kubenswrapper[4744]: I0106 16:28:15.098951 4744 scope.go:117] "RemoveContainer" containerID="988142daf64e81f40afde54d049e0ba87a0b8d44824ebec2048b209ae22d4911" Jan 06 16:28:15 crc kubenswrapper[4744]: I0106 16:28:15.731278 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae9105d-b113-4d86-9f24-733246388a8e" path="/var/lib/kubelet/pods/aae9105d-b113-4d86-9f24-733246388a8e/volumes" Jan 06 16:28:25 crc kubenswrapper[4744]: I0106 16:28:25.712105 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:28:25 crc kubenswrapper[4744]: E0106 16:28:25.713663 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:28:37 crc kubenswrapper[4744]: I0106 16:28:37.713155 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:28:37 crc kubenswrapper[4744]: E0106 16:28:37.714986 4744 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rm6lj_openshift-machine-config-operator(1639b17c-035d-4955-a3b6-ac33889e1c89)\"" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" Jan 06 16:28:48 crc kubenswrapper[4744]: I0106 16:28:48.712460 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:28:49 crc kubenswrapper[4744]: I0106 16:28:49.811007 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"f3cf439329878d4379eb3599a0b27195e28ebe65a9be7561b5e3ae2b1dbae537"} Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.227834 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c"] Jan 06 16:30:00 crc kubenswrapper[4744]: E0106 16:30:00.238810 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae9105d-b113-4d86-9f24-733246388a8e" containerName="extract-utilities" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.239156 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae9105d-b113-4d86-9f24-733246388a8e" containerName="extract-utilities" Jan 06 16:30:00 crc kubenswrapper[4744]: E0106 16:30:00.239217 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae9105d-b113-4d86-9f24-733246388a8e" containerName="extract-content" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.239230 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae9105d-b113-4d86-9f24-733246388a8e" containerName="extract-content" Jan 06 16:30:00 crc kubenswrapper[4744]: E0106 16:30:00.239291 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae9105d-b113-4d86-9f24-733246388a8e" containerName="registry-server" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.239301 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae9105d-b113-4d86-9f24-733246388a8e" containerName="registry-server" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.240494 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae9105d-b113-4d86-9f24-733246388a8e" containerName="registry-server" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.242271 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.246646 4744 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.246808 4744 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.260215 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c"] Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.373653 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2331a5b2-c9b5-4407-903b-f991b415b68d-secret-volume\") pod \"collect-profiles-29461950-5kl4c\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.374112 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjl67\" (UniqueName: \"kubernetes.io/projected/2331a5b2-c9b5-4407-903b-f991b415b68d-kube-api-access-rjl67\") pod \"collect-profiles-29461950-5kl4c\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.374245 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2331a5b2-c9b5-4407-903b-f991b415b68d-config-volume\") pod \"collect-profiles-29461950-5kl4c\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.476382 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjl67\" (UniqueName: \"kubernetes.io/projected/2331a5b2-c9b5-4407-903b-f991b415b68d-kube-api-access-rjl67\") pod \"collect-profiles-29461950-5kl4c\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.476538 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2331a5b2-c9b5-4407-903b-f991b415b68d-config-volume\") pod \"collect-profiles-29461950-5kl4c\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.477432 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2331a5b2-c9b5-4407-903b-f991b415b68d-secret-volume\") pod \"collect-profiles-29461950-5kl4c\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.477533 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2331a5b2-c9b5-4407-903b-f991b415b68d-config-volume\") pod \"collect-profiles-29461950-5kl4c\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.483645 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2331a5b2-c9b5-4407-903b-f991b415b68d-secret-volume\") pod \"collect-profiles-29461950-5kl4c\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.496114 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjl67\" (UniqueName: \"kubernetes.io/projected/2331a5b2-c9b5-4407-903b-f991b415b68d-kube-api-access-rjl67\") pod \"collect-profiles-29461950-5kl4c\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:00 crc kubenswrapper[4744]: I0106 16:30:00.585924 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:01 crc kubenswrapper[4744]: I0106 16:30:01.059669 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c"] Jan 06 16:30:01 crc kubenswrapper[4744]: I0106 16:30:01.693749 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" event={"ID":"2331a5b2-c9b5-4407-903b-f991b415b68d","Type":"ContainerStarted","Data":"0b9a4df3007781137774c901f802c2a1c814768e053c1e2984f650fe921be230"} Jan 06 16:30:01 crc kubenswrapper[4744]: I0106 16:30:01.694029 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" event={"ID":"2331a5b2-c9b5-4407-903b-f991b415b68d","Type":"ContainerStarted","Data":"88d3fa8be7f28da860059cb114b9e292251b9e7a48fef328bd55893b475afffb"} Jan 06 16:30:01 crc kubenswrapper[4744]: I0106 16:30:01.738123 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" podStartSLOduration=1.738104539 podStartE2EDuration="1.738104539s" podCreationTimestamp="2026-01-06 16:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-06 16:30:01.722320229 +0000 UTC m=+6798.349786587" watchObservedRunningTime="2026-01-06 16:30:01.738104539 +0000 UTC m=+6798.365570847" Jan 06 16:30:02 crc kubenswrapper[4744]: I0106 16:30:02.708299 4744 generic.go:334] "Generic (PLEG): container finished" podID="2331a5b2-c9b5-4407-903b-f991b415b68d" containerID="0b9a4df3007781137774c901f802c2a1c814768e053c1e2984f650fe921be230" exitCode=0 Jan 06 16:30:02 crc kubenswrapper[4744]: I0106 16:30:02.708369 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" event={"ID":"2331a5b2-c9b5-4407-903b-f991b415b68d","Type":"ContainerDied","Data":"0b9a4df3007781137774c901f802c2a1c814768e053c1e2984f650fe921be230"} Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.161827 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.273511 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjl67\" (UniqueName: \"kubernetes.io/projected/2331a5b2-c9b5-4407-903b-f991b415b68d-kube-api-access-rjl67\") pod \"2331a5b2-c9b5-4407-903b-f991b415b68d\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.273700 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2331a5b2-c9b5-4407-903b-f991b415b68d-secret-volume\") pod \"2331a5b2-c9b5-4407-903b-f991b415b68d\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.273791 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2331a5b2-c9b5-4407-903b-f991b415b68d-config-volume\") pod \"2331a5b2-c9b5-4407-903b-f991b415b68d\" (UID: \"2331a5b2-c9b5-4407-903b-f991b415b68d\") " Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.274605 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2331a5b2-c9b5-4407-903b-f991b415b68d-config-volume" (OuterVolumeSpecName: "config-volume") pod "2331a5b2-c9b5-4407-903b-f991b415b68d" (UID: "2331a5b2-c9b5-4407-903b-f991b415b68d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.281494 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2331a5b2-c9b5-4407-903b-f991b415b68d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2331a5b2-c9b5-4407-903b-f991b415b68d" (UID: "2331a5b2-c9b5-4407-903b-f991b415b68d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.281513 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2331a5b2-c9b5-4407-903b-f991b415b68d-kube-api-access-rjl67" (OuterVolumeSpecName: "kube-api-access-rjl67") pod "2331a5b2-c9b5-4407-903b-f991b415b68d" (UID: "2331a5b2-c9b5-4407-903b-f991b415b68d"). InnerVolumeSpecName "kube-api-access-rjl67". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.375945 4744 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2331a5b2-c9b5-4407-903b-f991b415b68d-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.375982 4744 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2331a5b2-c9b5-4407-903b-f991b415b68d-config-volume\") on node \"crc\" DevicePath \"\"" Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.375992 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjl67\" (UniqueName: \"kubernetes.io/projected/2331a5b2-c9b5-4407-903b-f991b415b68d-kube-api-access-rjl67\") on node \"crc\" DevicePath \"\"" Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.730007 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" event={"ID":"2331a5b2-c9b5-4407-903b-f991b415b68d","Type":"ContainerDied","Data":"88d3fa8be7f28da860059cb114b9e292251b9e7a48fef328bd55893b475afffb"} Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.730453 4744 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88d3fa8be7f28da860059cb114b9e292251b9e7a48fef328bd55893b475afffb" Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.730039 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29461950-5kl4c" Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.806358 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct"] Jan 06 16:30:04 crc kubenswrapper[4744]: I0106 16:30:04.818127 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29461905-vccct"] Jan 06 16:30:05 crc kubenswrapper[4744]: I0106 16:30:05.746131 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99512364-e502-4a80-9054-9af9f683744d" path="/var/lib/kubelet/pods/99512364-e502-4a80-9054-9af9f683744d/volumes" Jan 06 16:30:15 crc kubenswrapper[4744]: I0106 16:30:15.227649 4744 scope.go:117] "RemoveContainer" containerID="a9c69895f9613e32cac822451bb0e1741b2b7860db0511e0feab9e7714babeb5" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.197823 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2bzvr"] Jan 06 16:30:16 crc kubenswrapper[4744]: E0106 16:30:16.198772 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2331a5b2-c9b5-4407-903b-f991b415b68d" containerName="collect-profiles" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.198794 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="2331a5b2-c9b5-4407-903b-f991b415b68d" containerName="collect-profiles" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.199203 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="2331a5b2-c9b5-4407-903b-f991b415b68d" containerName="collect-profiles" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.201467 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.216791 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2bzvr"] Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.295580 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-utilities\") pod \"redhat-operators-2bzvr\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.296425 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qcn4\" (UniqueName: \"kubernetes.io/projected/619c228e-a44b-43cc-b86d-4e8a2eb7b331-kube-api-access-9qcn4\") pod \"redhat-operators-2bzvr\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.296682 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-catalog-content\") pod \"redhat-operators-2bzvr\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.399320 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-utilities\") pod \"redhat-operators-2bzvr\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.399362 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qcn4\" (UniqueName: \"kubernetes.io/projected/619c228e-a44b-43cc-b86d-4e8a2eb7b331-kube-api-access-9qcn4\") pod \"redhat-operators-2bzvr\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.399413 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-catalog-content\") pod \"redhat-operators-2bzvr\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.399949 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-catalog-content\") pod \"redhat-operators-2bzvr\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.400002 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-utilities\") pod \"redhat-operators-2bzvr\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.421017 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qcn4\" (UniqueName: \"kubernetes.io/projected/619c228e-a44b-43cc-b86d-4e8a2eb7b331-kube-api-access-9qcn4\") pod \"redhat-operators-2bzvr\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:16 crc kubenswrapper[4744]: I0106 16:30:16.523618 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:17 crc kubenswrapper[4744]: I0106 16:30:17.114751 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2bzvr"] Jan 06 16:30:17 crc kubenswrapper[4744]: I0106 16:30:17.914023 4744 generic.go:334] "Generic (PLEG): container finished" podID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerID="4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402" exitCode=0 Jan 06 16:30:17 crc kubenswrapper[4744]: I0106 16:30:17.914081 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bzvr" event={"ID":"619c228e-a44b-43cc-b86d-4e8a2eb7b331","Type":"ContainerDied","Data":"4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402"} Jan 06 16:30:17 crc kubenswrapper[4744]: I0106 16:30:17.914364 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bzvr" event={"ID":"619c228e-a44b-43cc-b86d-4e8a2eb7b331","Type":"ContainerStarted","Data":"eb4d3f58ed8f18dd9a8ecf2bd7397a44f561e75fca05303ba5cf62e1239a3ea4"} Jan 06 16:30:17 crc kubenswrapper[4744]: I0106 16:30:17.917766 4744 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 06 16:30:18 crc kubenswrapper[4744]: I0106 16:30:18.932117 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bzvr" event={"ID":"619c228e-a44b-43cc-b86d-4e8a2eb7b331","Type":"ContainerStarted","Data":"f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955"} Jan 06 16:30:22 crc kubenswrapper[4744]: I0106 16:30:22.986430 4744 generic.go:334] "Generic (PLEG): container finished" podID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerID="f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955" exitCode=0 Jan 06 16:30:22 crc kubenswrapper[4744]: I0106 16:30:22.986495 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bzvr" event={"ID":"619c228e-a44b-43cc-b86d-4e8a2eb7b331","Type":"ContainerDied","Data":"f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955"} Jan 06 16:30:23 crc kubenswrapper[4744]: I0106 16:30:23.998402 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bzvr" event={"ID":"619c228e-a44b-43cc-b86d-4e8a2eb7b331","Type":"ContainerStarted","Data":"b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac"} Jan 06 16:30:24 crc kubenswrapper[4744]: I0106 16:30:24.022752 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2bzvr" podStartSLOduration=2.50091675 podStartE2EDuration="8.022732768s" podCreationTimestamp="2026-01-06 16:30:16 +0000 UTC" firstStartedPulling="2026-01-06 16:30:17.915773384 +0000 UTC m=+6814.543239702" lastFinishedPulling="2026-01-06 16:30:23.437589402 +0000 UTC m=+6820.065055720" observedRunningTime="2026-01-06 16:30:24.020075087 +0000 UTC m=+6820.647541405" watchObservedRunningTime="2026-01-06 16:30:24.022732768 +0000 UTC m=+6820.650199086" Jan 06 16:30:26 crc kubenswrapper[4744]: I0106 16:30:26.524362 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:26 crc kubenswrapper[4744]: I0106 16:30:26.525014 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:27 crc kubenswrapper[4744]: I0106 16:30:27.592761 4744 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2bzvr" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerName="registry-server" probeResult="failure" output=< Jan 06 16:30:27 crc kubenswrapper[4744]: timeout: failed to connect service ":50051" within 1s Jan 06 16:30:27 crc kubenswrapper[4744]: > Jan 06 16:30:36 crc kubenswrapper[4744]: I0106 16:30:36.603296 4744 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:36 crc kubenswrapper[4744]: I0106 16:30:36.677206 4744 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:36 crc kubenswrapper[4744]: I0106 16:30:36.863683 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2bzvr"] Jan 06 16:30:38 crc kubenswrapper[4744]: I0106 16:30:38.160240 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2bzvr" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerName="registry-server" containerID="cri-o://b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac" gracePeriod=2 Jan 06 16:30:38 crc kubenswrapper[4744]: I0106 16:30:38.774302 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:38 crc kubenswrapper[4744]: I0106 16:30:38.905371 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qcn4\" (UniqueName: \"kubernetes.io/projected/619c228e-a44b-43cc-b86d-4e8a2eb7b331-kube-api-access-9qcn4\") pod \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " Jan 06 16:30:38 crc kubenswrapper[4744]: I0106 16:30:38.905750 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-utilities\") pod \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " Jan 06 16:30:38 crc kubenswrapper[4744]: I0106 16:30:38.906142 4744 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-catalog-content\") pod \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\" (UID: \"619c228e-a44b-43cc-b86d-4e8a2eb7b331\") " Jan 06 16:30:38 crc kubenswrapper[4744]: I0106 16:30:38.907078 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-utilities" (OuterVolumeSpecName: "utilities") pod "619c228e-a44b-43cc-b86d-4e8a2eb7b331" (UID: "619c228e-a44b-43cc-b86d-4e8a2eb7b331"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:30:38 crc kubenswrapper[4744]: I0106 16:30:38.907843 4744 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-utilities\") on node \"crc\" DevicePath \"\"" Jan 06 16:30:38 crc kubenswrapper[4744]: I0106 16:30:38.915146 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/619c228e-a44b-43cc-b86d-4e8a2eb7b331-kube-api-access-9qcn4" (OuterVolumeSpecName: "kube-api-access-9qcn4") pod "619c228e-a44b-43cc-b86d-4e8a2eb7b331" (UID: "619c228e-a44b-43cc-b86d-4e8a2eb7b331"). InnerVolumeSpecName "kube-api-access-9qcn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.005140 4744 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "619c228e-a44b-43cc-b86d-4e8a2eb7b331" (UID: "619c228e-a44b-43cc-b86d-4e8a2eb7b331"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.011189 4744 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qcn4\" (UniqueName: \"kubernetes.io/projected/619c228e-a44b-43cc-b86d-4e8a2eb7b331-kube-api-access-9qcn4\") on node \"crc\" DevicePath \"\"" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.011257 4744 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/619c228e-a44b-43cc-b86d-4e8a2eb7b331-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.180699 4744 generic.go:334] "Generic (PLEG): container finished" podID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerID="b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac" exitCode=0 Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.180815 4744 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2bzvr" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.180800 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bzvr" event={"ID":"619c228e-a44b-43cc-b86d-4e8a2eb7b331","Type":"ContainerDied","Data":"b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac"} Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.181023 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2bzvr" event={"ID":"619c228e-a44b-43cc-b86d-4e8a2eb7b331","Type":"ContainerDied","Data":"eb4d3f58ed8f18dd9a8ecf2bd7397a44f561e75fca05303ba5cf62e1239a3ea4"} Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.181073 4744 scope.go:117] "RemoveContainer" containerID="b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.211769 4744 scope.go:117] "RemoveContainer" containerID="f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.244873 4744 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2bzvr"] Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.256413 4744 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2bzvr"] Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.277841 4744 scope.go:117] "RemoveContainer" containerID="4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.332599 4744 scope.go:117] "RemoveContainer" containerID="b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac" Jan 06 16:30:39 crc kubenswrapper[4744]: E0106 16:30:39.333313 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac\": container with ID starting with b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac not found: ID does not exist" containerID="b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.333371 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac"} err="failed to get container status \"b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac\": rpc error: code = NotFound desc = could not find container \"b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac\": container with ID starting with b04c777965a41a36c74152d35fec456fe5cc7182be950a17f042a9fceb062cac not found: ID does not exist" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.333411 4744 scope.go:117] "RemoveContainer" containerID="f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955" Jan 06 16:30:39 crc kubenswrapper[4744]: E0106 16:30:39.334248 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955\": container with ID starting with f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955 not found: ID does not exist" containerID="f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.334330 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955"} err="failed to get container status \"f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955\": rpc error: code = NotFound desc = could not find container \"f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955\": container with ID starting with f972d5d0997333d4a12c0a2744a1531ffc4ac4417bf12733806b0851fe555955 not found: ID does not exist" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.334358 4744 scope.go:117] "RemoveContainer" containerID="4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402" Jan 06 16:30:39 crc kubenswrapper[4744]: E0106 16:30:39.334773 4744 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402\": container with ID starting with 4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402 not found: ID does not exist" containerID="4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.334817 4744 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402"} err="failed to get container status \"4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402\": rpc error: code = NotFound desc = could not find container \"4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402\": container with ID starting with 4b4ddc9e888187ee27bbc77954231b791a344827960eb9cdcb2af984e6c33402 not found: ID does not exist" Jan 06 16:30:39 crc kubenswrapper[4744]: I0106 16:30:39.729061 4744 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" path="/var/lib/kubelet/pods/619c228e-a44b-43cc-b86d-4e8a2eb7b331/volumes" Jan 06 16:31:14 crc kubenswrapper[4744]: I0106 16:31:14.424060 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:31:14 crc kubenswrapper[4744]: I0106 16:31:14.424623 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:31:44 crc kubenswrapper[4744]: I0106 16:31:44.423935 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:31:44 crc kubenswrapper[4744]: I0106 16:31:44.424665 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:32:14 crc kubenswrapper[4744]: I0106 16:32:14.423559 4744 patch_prober.go:28] interesting pod/machine-config-daemon-rm6lj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 06 16:32:14 crc kubenswrapper[4744]: I0106 16:32:14.424253 4744 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 06 16:32:14 crc kubenswrapper[4744]: I0106 16:32:14.424324 4744 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" Jan 06 16:32:14 crc kubenswrapper[4744]: I0106 16:32:14.425677 4744 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3cf439329878d4379eb3599a0b27195e28ebe65a9be7561b5e3ae2b1dbae537"} pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 06 16:32:14 crc kubenswrapper[4744]: I0106 16:32:14.425944 4744 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" podUID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerName="machine-config-daemon" containerID="cri-o://f3cf439329878d4379eb3599a0b27195e28ebe65a9be7561b5e3ae2b1dbae537" gracePeriod=600 Jan 06 16:32:15 crc kubenswrapper[4744]: I0106 16:32:15.521366 4744 generic.go:334] "Generic (PLEG): container finished" podID="1639b17c-035d-4955-a3b6-ac33889e1c89" containerID="f3cf439329878d4379eb3599a0b27195e28ebe65a9be7561b5e3ae2b1dbae537" exitCode=0 Jan 06 16:32:15 crc kubenswrapper[4744]: I0106 16:32:15.521460 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerDied","Data":"f3cf439329878d4379eb3599a0b27195e28ebe65a9be7561b5e3ae2b1dbae537"} Jan 06 16:32:15 crc kubenswrapper[4744]: I0106 16:32:15.521711 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rm6lj" event={"ID":"1639b17c-035d-4955-a3b6-ac33889e1c89","Type":"ContainerStarted","Data":"947d2762c06529fed1f79b376a9d9e9c5d8697a3edf0df88eee30f649925757e"} Jan 06 16:32:15 crc kubenswrapper[4744]: I0106 16:32:15.521733 4744 scope.go:117] "RemoveContainer" containerID="c45bb894dc20dd80910c76df78695ce3dba0ed2bc8808f73b85d00f0b8768e2f" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.285613 4744 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gkcxf"] Jan 06 16:33:13 crc kubenswrapper[4744]: E0106 16:33:13.286606 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerName="extract-utilities" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.286622 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerName="extract-utilities" Jan 06 16:33:13 crc kubenswrapper[4744]: E0106 16:33:13.286661 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerName="registry-server" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.286670 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerName="registry-server" Jan 06 16:33:13 crc kubenswrapper[4744]: E0106 16:33:13.286688 4744 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerName="extract-content" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.286696 4744 state_mem.go:107] "Deleted CPUSet assignment" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerName="extract-content" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.287040 4744 memory_manager.go:354] "RemoveStaleState removing state" podUID="619c228e-a44b-43cc-b86d-4e8a2eb7b331" containerName="registry-server" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.289194 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.296511 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gkcxf"] Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.446599 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b9818a-8a0f-416d-bd64-25533893f485-catalog-content\") pod \"certified-operators-gkcxf\" (UID: \"f5b9818a-8a0f-416d-bd64-25533893f485\") " pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.446929 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b9818a-8a0f-416d-bd64-25533893f485-utilities\") pod \"certified-operators-gkcxf\" (UID: \"f5b9818a-8a0f-416d-bd64-25533893f485\") " pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.447291 4744 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfs9w\" (UniqueName: \"kubernetes.io/projected/f5b9818a-8a0f-416d-bd64-25533893f485-kube-api-access-pfs9w\") pod \"certified-operators-gkcxf\" (UID: \"f5b9818a-8a0f-416d-bd64-25533893f485\") " pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.549753 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b9818a-8a0f-416d-bd64-25533893f485-catalog-content\") pod \"certified-operators-gkcxf\" (UID: \"f5b9818a-8a0f-416d-bd64-25533893f485\") " pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.549880 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b9818a-8a0f-416d-bd64-25533893f485-utilities\") pod \"certified-operators-gkcxf\" (UID: \"f5b9818a-8a0f-416d-bd64-25533893f485\") " pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.549990 4744 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfs9w\" (UniqueName: \"kubernetes.io/projected/f5b9818a-8a0f-416d-bd64-25533893f485-kube-api-access-pfs9w\") pod \"certified-operators-gkcxf\" (UID: \"f5b9818a-8a0f-416d-bd64-25533893f485\") " pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.550888 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5b9818a-8a0f-416d-bd64-25533893f485-catalog-content\") pod \"certified-operators-gkcxf\" (UID: \"f5b9818a-8a0f-416d-bd64-25533893f485\") " pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.550929 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5b9818a-8a0f-416d-bd64-25533893f485-utilities\") pod \"certified-operators-gkcxf\" (UID: \"f5b9818a-8a0f-416d-bd64-25533893f485\") " pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.576958 4744 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfs9w\" (UniqueName: \"kubernetes.io/projected/f5b9818a-8a0f-416d-bd64-25533893f485-kube-api-access-pfs9w\") pod \"certified-operators-gkcxf\" (UID: \"f5b9818a-8a0f-416d-bd64-25533893f485\") " pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:13 crc kubenswrapper[4744]: I0106 16:33:13.606622 4744 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gkcxf" Jan 06 16:33:14 crc kubenswrapper[4744]: I0106 16:33:14.126134 4744 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gkcxf"] Jan 06 16:33:14 crc kubenswrapper[4744]: I0106 16:33:14.377324 4744 generic.go:334] "Generic (PLEG): container finished" podID="f5b9818a-8a0f-416d-bd64-25533893f485" containerID="e6b3bd079fbb8233cc5ab4c6c11e7eed1632ce5db2dc69c45747fb81d9f400a3" exitCode=0 Jan 06 16:33:14 crc kubenswrapper[4744]: I0106 16:33:14.377384 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkcxf" event={"ID":"f5b9818a-8a0f-416d-bd64-25533893f485","Type":"ContainerDied","Data":"e6b3bd079fbb8233cc5ab4c6c11e7eed1632ce5db2dc69c45747fb81d9f400a3"} Jan 06 16:33:14 crc kubenswrapper[4744]: I0106 16:33:14.377732 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkcxf" event={"ID":"f5b9818a-8a0f-416d-bd64-25533893f485","Type":"ContainerStarted","Data":"7df4371683fb9245a6548f1ab1a370912aed101b34a75c7d369553a5b44f6dc3"} Jan 06 16:33:16 crc kubenswrapper[4744]: I0106 16:33:16.404748 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkcxf" event={"ID":"f5b9818a-8a0f-416d-bd64-25533893f485","Type":"ContainerStarted","Data":"4a6f942cc4457d524188b1a38a1d4e7f560875f55fbd0bb6870666bdff72e5b8"} Jan 06 16:33:18 crc kubenswrapper[4744]: I0106 16:33:18.441025 4744 generic.go:334] "Generic (PLEG): container finished" podID="f5b9818a-8a0f-416d-bd64-25533893f485" containerID="4a6f942cc4457d524188b1a38a1d4e7f560875f55fbd0bb6870666bdff72e5b8" exitCode=0 Jan 06 16:33:18 crc kubenswrapper[4744]: I0106 16:33:18.441208 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkcxf" event={"ID":"f5b9818a-8a0f-416d-bd64-25533893f485","Type":"ContainerDied","Data":"4a6f942cc4457d524188b1a38a1d4e7f560875f55fbd0bb6870666bdff72e5b8"} Jan 06 16:33:19 crc kubenswrapper[4744]: I0106 16:33:19.456397 4744 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gkcxf" event={"ID":"f5b9818a-8a0f-416d-bd64-25533893f485","Type":"ContainerStarted","Data":"f5ab4025ba64d13ebf01ff93ed37943cbb2c4f6ddc7a95e5f8bebd2d8460cc79"} Jan 06 16:33:19 crc kubenswrapper[4744]: I0106 16:33:19.482016 4744 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gkcxf" podStartSLOduration=1.871669158 podStartE2EDuration="6.481993344s" podCreationTimestamp="2026-01-06 16:33:13 +0000 UTC" firstStartedPulling="2026-01-06 16:33:14.379886267 +0000 UTC m=+6991.007352575" lastFinishedPulling="2026-01-06 16:33:18.990210403 +0000 UTC m=+6995.617676761" observedRunningTime="2026-01-06 16:33:19.476373844 +0000 UTC m=+6996.103840212" watchObservedRunningTime="2026-01-06 16:33:19.481993344 +0000 UTC m=+6996.109459672" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515127234527024455 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015127234530017364 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015127216431016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015127216431015457 5ustar corecore